var/home/core/zuul-output/0000755000175000017500000000000015111112212014510 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111123032015456 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004774466715111123023017713 0ustar rootrootNov 24 17:26:46 crc systemd[1]: Starting Kubernetes Kubelet... Nov 24 17:26:46 crc restorecon[4690]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:26:47 crc restorecon[4690]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:26:47 crc restorecon[4690]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 24 17:26:47 crc kubenswrapper[4808]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 17:26:47 crc kubenswrapper[4808]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 24 17:26:47 crc kubenswrapper[4808]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 17:26:47 crc kubenswrapper[4808]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 17:26:47 crc kubenswrapper[4808]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 24 17:26:47 crc kubenswrapper[4808]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.922255 4808 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931549 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931587 4808 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931592 4808 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931596 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931602 4808 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931609 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931613 4808 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931616 4808 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931622 4808 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931628 4808 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931631 4808 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931636 4808 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931640 4808 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931644 4808 feature_gate.go:330] unrecognized feature gate: Example Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931649 4808 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931653 4808 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931658 4808 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931662 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931667 4808 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931671 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931675 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931679 4808 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931687 4808 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931728 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931732 4808 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931736 4808 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931740 4808 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931744 4808 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931748 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931752 4808 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931756 4808 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931760 4808 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931763 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931769 4808 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931774 4808 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931778 4808 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931782 4808 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931786 4808 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931790 4808 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931794 4808 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931799 4808 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931804 4808 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931808 4808 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931812 4808 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931817 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931821 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931825 4808 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931828 4808 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931832 4808 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931836 4808 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931839 4808 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931845 4808 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931849 4808 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931853 4808 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931857 4808 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931861 4808 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931865 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931869 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931872 4808 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931875 4808 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931879 4808 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931882 4808 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931886 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931890 4808 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931893 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931897 4808 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931900 4808 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931904 4808 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931908 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931912 4808 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.931915 4808 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932008 4808 flags.go:64] FLAG: --address="0.0.0.0" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932034 4808 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932042 4808 flags.go:64] FLAG: --anonymous-auth="true" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932049 4808 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932055 4808 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932060 4808 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932066 4808 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932072 4808 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932076 4808 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932080 4808 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932085 4808 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932090 4808 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932096 4808 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932100 4808 flags.go:64] FLAG: --cgroup-root="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932104 4808 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932109 4808 flags.go:64] FLAG: --client-ca-file="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932113 4808 flags.go:64] FLAG: --cloud-config="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932117 4808 flags.go:64] FLAG: --cloud-provider="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932121 4808 flags.go:64] FLAG: --cluster-dns="[]" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932127 4808 flags.go:64] FLAG: --cluster-domain="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932131 4808 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932135 4808 flags.go:64] FLAG: --config-dir="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932140 4808 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932145 4808 flags.go:64] FLAG: --container-log-max-files="5" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932151 4808 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932156 4808 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932161 4808 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932166 4808 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932171 4808 flags.go:64] FLAG: --contention-profiling="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932175 4808 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932180 4808 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932185 4808 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932189 4808 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932196 4808 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932203 4808 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932207 4808 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932212 4808 flags.go:64] FLAG: --enable-load-reader="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932216 4808 flags.go:64] FLAG: --enable-server="true" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932221 4808 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932228 4808 flags.go:64] FLAG: --event-burst="100" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932232 4808 flags.go:64] FLAG: --event-qps="50" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932237 4808 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932242 4808 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932247 4808 flags.go:64] FLAG: --eviction-hard="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932259 4808 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932264 4808 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932269 4808 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932274 4808 flags.go:64] FLAG: --eviction-soft="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932278 4808 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932282 4808 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932286 4808 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932290 4808 flags.go:64] FLAG: --experimental-mounter-path="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932295 4808 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932299 4808 flags.go:64] FLAG: --fail-swap-on="true" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932303 4808 flags.go:64] FLAG: --feature-gates="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932308 4808 flags.go:64] FLAG: --file-check-frequency="20s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932312 4808 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932316 4808 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932321 4808 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932325 4808 flags.go:64] FLAG: --healthz-port="10248" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932329 4808 flags.go:64] FLAG: --help="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932333 4808 flags.go:64] FLAG: --hostname-override="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932337 4808 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932341 4808 flags.go:64] FLAG: --http-check-frequency="20s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932345 4808 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932350 4808 flags.go:64] FLAG: --image-credential-provider-config="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932353 4808 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932358 4808 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932362 4808 flags.go:64] FLAG: --image-service-endpoint="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932365 4808 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932370 4808 flags.go:64] FLAG: --kube-api-burst="100" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932374 4808 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932379 4808 flags.go:64] FLAG: --kube-api-qps="50" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932384 4808 flags.go:64] FLAG: --kube-reserved="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932388 4808 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932393 4808 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932398 4808 flags.go:64] FLAG: --kubelet-cgroups="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932402 4808 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932407 4808 flags.go:64] FLAG: --lock-file="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932412 4808 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932416 4808 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932421 4808 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932427 4808 flags.go:64] FLAG: --log-json-split-stream="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932432 4808 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932436 4808 flags.go:64] FLAG: --log-text-split-stream="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932441 4808 flags.go:64] FLAG: --logging-format="text" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932445 4808 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932450 4808 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932454 4808 flags.go:64] FLAG: --manifest-url="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932458 4808 flags.go:64] FLAG: --manifest-url-header="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932466 4808 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932471 4808 flags.go:64] FLAG: --max-open-files="1000000" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932478 4808 flags.go:64] FLAG: --max-pods="110" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932483 4808 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932488 4808 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932493 4808 flags.go:64] FLAG: --memory-manager-policy="None" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932500 4808 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932505 4808 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932510 4808 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932516 4808 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932531 4808 flags.go:64] FLAG: --node-status-max-images="50" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932536 4808 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932542 4808 flags.go:64] FLAG: --oom-score-adj="-999" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932548 4808 flags.go:64] FLAG: --pod-cidr="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932554 4808 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932572 4808 flags.go:64] FLAG: --pod-manifest-path="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932578 4808 flags.go:64] FLAG: --pod-max-pids="-1" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932583 4808 flags.go:64] FLAG: --pods-per-core="0" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932588 4808 flags.go:64] FLAG: --port="10250" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932594 4808 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932598 4808 flags.go:64] FLAG: --provider-id="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932602 4808 flags.go:64] FLAG: --qos-reserved="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932606 4808 flags.go:64] FLAG: --read-only-port="10255" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932610 4808 flags.go:64] FLAG: --register-node="true" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932614 4808 flags.go:64] FLAG: --register-schedulable="true" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932619 4808 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932627 4808 flags.go:64] FLAG: --registry-burst="10" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932632 4808 flags.go:64] FLAG: --registry-qps="5" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932637 4808 flags.go:64] FLAG: --reserved-cpus="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932643 4808 flags.go:64] FLAG: --reserved-memory="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932650 4808 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932654 4808 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932659 4808 flags.go:64] FLAG: --rotate-certificates="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932663 4808 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932667 4808 flags.go:64] FLAG: --runonce="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932671 4808 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932676 4808 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932680 4808 flags.go:64] FLAG: --seccomp-default="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932684 4808 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932688 4808 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932693 4808 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932697 4808 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932702 4808 flags.go:64] FLAG: --storage-driver-password="root" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932706 4808 flags.go:64] FLAG: --storage-driver-secure="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932710 4808 flags.go:64] FLAG: --storage-driver-table="stats" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932714 4808 flags.go:64] FLAG: --storage-driver-user="root" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932718 4808 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932724 4808 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932729 4808 flags.go:64] FLAG: --system-cgroups="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932733 4808 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932741 4808 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932750 4808 flags.go:64] FLAG: --tls-cert-file="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932755 4808 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932760 4808 flags.go:64] FLAG: --tls-min-version="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932765 4808 flags.go:64] FLAG: --tls-private-key-file="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932770 4808 flags.go:64] FLAG: --topology-manager-policy="none" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932776 4808 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932781 4808 flags.go:64] FLAG: --topology-manager-scope="container" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932786 4808 flags.go:64] FLAG: --v="2" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932794 4808 flags.go:64] FLAG: --version="false" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932801 4808 flags.go:64] FLAG: --vmodule="" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932806 4808 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.932811 4808 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.932946 4808 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.932952 4808 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.932958 4808 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.932963 4808 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.932969 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.932974 4808 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.932979 4808 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.932984 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.932988 4808 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.932993 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.932997 4808 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933002 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933006 4808 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933011 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933037 4808 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933042 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933048 4808 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933052 4808 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933056 4808 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933060 4808 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933065 4808 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933069 4808 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933074 4808 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933078 4808 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933082 4808 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933086 4808 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933090 4808 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933094 4808 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933098 4808 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933101 4808 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933105 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933109 4808 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933112 4808 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933116 4808 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933120 4808 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933123 4808 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933127 4808 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933131 4808 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933137 4808 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933141 4808 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933145 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933149 4808 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933153 4808 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933350 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933354 4808 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933358 4808 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933361 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933365 4808 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933370 4808 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933373 4808 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933377 4808 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933381 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933386 4808 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933389 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933394 4808 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933399 4808 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933403 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933407 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933411 4808 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933415 4808 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933419 4808 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933423 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933426 4808 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933430 4808 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933434 4808 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933438 4808 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933442 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933445 4808 feature_gate.go:330] unrecognized feature gate: Example Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933449 4808 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933453 4808 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.933457 4808 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.934491 4808 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.945129 4808 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.945413 4808 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945624 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945658 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945668 4808 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945679 4808 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945688 4808 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945697 4808 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945705 4808 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945716 4808 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945724 4808 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945733 4808 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945742 4808 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945751 4808 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945759 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945767 4808 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945777 4808 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945787 4808 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945799 4808 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945812 4808 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945828 4808 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945847 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945860 4808 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945871 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945880 4808 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945888 4808 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945896 4808 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945905 4808 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945913 4808 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945921 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945929 4808 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945936 4808 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945945 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945956 4808 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945967 4808 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945976 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945988 4808 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.945998 4808 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946006 4808 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946046 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946057 4808 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946068 4808 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946079 4808 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946089 4808 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946101 4808 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946111 4808 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946124 4808 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946135 4808 feature_gate.go:330] unrecognized feature gate: Example Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946145 4808 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946155 4808 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946165 4808 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946178 4808 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946191 4808 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946202 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946212 4808 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946220 4808 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946228 4808 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946237 4808 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946245 4808 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946254 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946263 4808 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946271 4808 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946279 4808 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946288 4808 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946296 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946305 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946312 4808 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946322 4808 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946333 4808 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946342 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946350 4808 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946359 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946371 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.946386 4808 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946670 4808 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946685 4808 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946700 4808 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946713 4808 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946729 4808 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946743 4808 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946757 4808 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946770 4808 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946782 4808 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946792 4808 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946801 4808 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946810 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946819 4808 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946827 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946836 4808 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946844 4808 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946853 4808 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946861 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946869 4808 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946877 4808 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946885 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946893 4808 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946901 4808 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946910 4808 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946917 4808 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946925 4808 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946933 4808 feature_gate.go:330] unrecognized feature gate: Example Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946942 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946951 4808 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946961 4808 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946971 4808 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946981 4808 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.946992 4808 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947001 4808 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947047 4808 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947059 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947070 4808 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947080 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947090 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947100 4808 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947111 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947121 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947131 4808 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947141 4808 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947151 4808 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947163 4808 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947173 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947183 4808 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947194 4808 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947204 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947213 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947222 4808 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947230 4808 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947238 4808 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947247 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947255 4808 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947264 4808 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947271 4808 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947279 4808 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947286 4808 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947294 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947302 4808 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947310 4808 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947317 4808 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947328 4808 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947337 4808 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947347 4808 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947355 4808 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947364 4808 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947372 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 17:26:47 crc kubenswrapper[4808]: W1124 17:26:47.947381 4808 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.947395 4808 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.947775 4808 server.go:940] "Client rotation is on, will bootstrap in background" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.954819 4808 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.955000 4808 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.957563 4808 server.go:997] "Starting client certificate rotation" Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.957615 4808 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.957833 4808 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-14 21:30:47.462715161 +0000 UTC Nov 24 17:26:47 crc kubenswrapper[4808]: I1124 17:26:47.957941 4808 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 484h3m59.504777638s for next certificate rotation Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.016703 4808 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.020962 4808 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.187778 4808 log.go:25] "Validated CRI v1 runtime API" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.221230 4808 log.go:25] "Validated CRI v1 image API" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.223365 4808 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.236839 4808 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-24-17-22-29-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.236913 4808 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:44 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.259168 4808 manager.go:217] Machine: {Timestamp:2025-11-24 17:26:48.257103761 +0000 UTC m=+0.854771613 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:c3e5358f-bf99-4bb9-b876-ca5a1061a1cf BootID:8a5edfdd-9e82-47ae-ad4b-d584fa1b687e Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:44 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:0e:6a:6c Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:0e:6a:6c Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:6e:55:5d Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:a9:27:8d Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:75:41:12 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:c5:00:67 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:d6:27:62:cf:f5:f4 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:c6:68:e5:ec:39:1f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.259469 4808 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.259689 4808 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.261032 4808 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.261322 4808 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.261367 4808 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.261632 4808 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.261645 4808 container_manager_linux.go:303] "Creating device plugin manager" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.262327 4808 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.262365 4808 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.262704 4808 state_mem.go:36] "Initialized new in-memory state store" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.262812 4808 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.267099 4808 kubelet.go:418] "Attempting to sync node with API server" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.267130 4808 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.267179 4808 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.267201 4808 kubelet.go:324] "Adding apiserver pod source" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.267218 4808 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 17:26:48 crc kubenswrapper[4808]: W1124 17:26:48.275596 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.171:6443: connect: connection refused Nov 24 17:26:48 crc kubenswrapper[4808]: W1124 17:26:48.275572 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.171:6443: connect: connection refused Nov 24 17:26:48 crc kubenswrapper[4808]: E1124 17:26:48.275719 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.171:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:26:48 crc kubenswrapper[4808]: E1124 17:26:48.275753 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.171:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.276075 4808 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.276957 4808 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.281744 4808 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.283535 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.283562 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.283570 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.283577 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.283590 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.283600 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.283608 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.283620 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.283637 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.283662 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.283677 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.283685 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.283705 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.284308 4808 server.go:1280] "Started kubelet" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.284793 4808 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.285355 4808 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.286190 4808 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.286333 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.286360 4808 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.286444 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 00:14:26.279218442 +0000 UTC Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.286517 4808 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.286547 4808 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.286706 4808 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 24 17:26:48 crc kubenswrapper[4808]: E1124 17:26:48.286529 4808 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.287099 4808 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.171:6443: connect: connection refused Nov 24 17:26:48 crc kubenswrapper[4808]: W1124 17:26:48.287180 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.171:6443: connect: connection refused Nov 24 17:26:48 crc systemd[1]: Started Kubernetes Kubelet. Nov 24 17:26:48 crc kubenswrapper[4808]: E1124 17:26:48.287240 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.171:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.287472 4808 factory.go:55] Registering systemd factory Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.287488 4808 factory.go:221] Registration of the systemd container factory successfully Nov 24 17:26:48 crc kubenswrapper[4808]: E1124 17:26:48.293041 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.171:6443: connect: connection refused" interval="200ms" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.295341 4808 factory.go:153] Registering CRI-O factory Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.295417 4808 factory.go:221] Registration of the crio container factory successfully Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.295515 4808 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.295622 4808 factory.go:103] Registering Raw factory Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.295689 4808 manager.go:1196] Started watching for new ooms in manager Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.296622 4808 manager.go:319] Starting recovery of all containers Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.299270 4808 server.go:460] "Adding debug handlers to kubelet server" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.302679 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.302831 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.302907 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303106 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303183 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303240 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303302 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: E1124 17:26:48.299219 4808 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.171:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b01640db91f34 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 17:26:48.284274484 +0000 UTC m=+0.881942286,LastTimestamp:2025-11-24 17:26:48.284274484 +0000 UTC m=+0.881942286,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303366 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303515 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303579 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303629 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303653 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303678 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303705 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303722 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303743 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303795 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303816 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303835 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303854 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303872 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303892 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303909 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303952 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303976 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.303998 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304037 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304059 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304076 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304119 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304155 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304175 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304199 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304217 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304235 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304279 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304300 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304317 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304335 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304351 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304370 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304392 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304408 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304455 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304474 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304491 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304506 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304525 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304543 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304563 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304611 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304667 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304695 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304716 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304734 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304778 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304795 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304815 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304834 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304851 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304871 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304887 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304928 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304947 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304962 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304979 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.304996 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305046 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305071 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305090 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305106 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305123 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305138 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305157 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305172 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305217 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305236 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305256 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305273 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305290 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305308 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305327 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.305898 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.306642 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307574 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307590 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307605 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307625 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307638 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307709 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307721 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307736 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307747 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307760 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307772 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307803 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307814 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307827 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307838 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307848 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307860 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307873 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307887 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307900 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307919 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307934 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307946 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307959 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307972 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307986 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.307998 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308010 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308038 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308051 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308067 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308079 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308093 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308106 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308117 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308129 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308154 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308168 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308181 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308194 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308210 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308222 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308233 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308244 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308256 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308268 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308282 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308294 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308308 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308319 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308331 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308342 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308354 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308365 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308376 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308388 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308408 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308423 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308439 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308450 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308459 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308471 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308482 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308495 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308507 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308517 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308528 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308539 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308552 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308567 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308578 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308588 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308601 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308615 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308627 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308636 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308646 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308658 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308668 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308678 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308689 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308701 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308711 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308721 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308733 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308746 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308756 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308766 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308779 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308797 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308844 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308857 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308868 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308879 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308892 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308902 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308914 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308925 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308935 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308947 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308958 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308977 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308987 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.308995 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.309007 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.309276 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310359 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310370 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310383 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310397 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310411 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310422 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310435 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310445 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310455 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310464 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310474 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310486 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310495 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310504 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310514 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.310527 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.313856 4808 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.313971 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.314145 4808 reconstruct.go:97] "Volume reconstruction finished" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.314215 4808 reconciler.go:26] "Reconciler: start to sync state" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.314875 4808 manager.go:324] Recovery completed Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.323489 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.328530 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.328603 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.328666 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.329986 4808 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.330045 4808 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.330084 4808 state_mem.go:36] "Initialized new in-memory state store" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.343603 4808 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.345974 4808 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.346045 4808 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.346091 4808 kubelet.go:2335] "Starting kubelet main sync loop" Nov 24 17:26:48 crc kubenswrapper[4808]: E1124 17:26:48.346180 4808 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 17:26:48 crc kubenswrapper[4808]: W1124 17:26:48.347507 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.171:6443: connect: connection refused Nov 24 17:26:48 crc kubenswrapper[4808]: E1124 17:26:48.347574 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.171:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.348876 4808 policy_none.go:49] "None policy: Start" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.349570 4808 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.349606 4808 state_mem.go:35] "Initializing new in-memory state store" Nov 24 17:26:48 crc kubenswrapper[4808]: E1124 17:26:48.386826 4808 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.404516 4808 manager.go:334] "Starting Device Plugin manager" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.404599 4808 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.404617 4808 server.go:79] "Starting device plugin registration server" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.405241 4808 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.405261 4808 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.405396 4808 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.405524 4808 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.405539 4808 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 17:26:48 crc kubenswrapper[4808]: E1124 17:26:48.417863 4808 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.447222 4808 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.447384 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.449271 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.449317 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.449331 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.449540 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.449909 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.449971 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.450750 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.450785 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.450802 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.450901 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.451058 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.451107 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.451237 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.451285 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.451298 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.451679 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.451708 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.451721 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.451807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.451826 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.451839 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.451842 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.451938 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.451977 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.452554 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.452580 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.452594 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.452724 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.452945 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.452994 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.453009 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.453058 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.453078 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.453687 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.453724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.453742 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.453936 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.453955 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.453963 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.453992 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.454058 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.455041 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.455068 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.455081 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:48 crc kubenswrapper[4808]: E1124 17:26:48.494214 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.171:6443: connect: connection refused" interval="400ms" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.505721 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.507336 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.507466 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.507489 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.507529 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 17:26:48 crc kubenswrapper[4808]: E1124 17:26:48.508273 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.171:6443: connect: connection refused" node="crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.516498 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.516559 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.516583 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.516623 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.516642 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.516661 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.516731 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.516894 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.516979 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.517085 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.517156 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.517405 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.517470 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.517526 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.517556 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618337 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618415 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618442 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618469 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618492 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618512 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618533 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618552 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618573 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618590 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618605 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618620 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618637 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618653 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618669 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618721 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.618984 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.619088 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.619183 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.619240 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.619245 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.619281 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.619320 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.619318 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.619369 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.619394 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.619435 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.619442 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.619464 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.619532 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.709284 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.711491 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.711590 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.711613 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.711658 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 17:26:48 crc kubenswrapper[4808]: E1124 17:26:48.712409 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.171:6443: connect: connection refused" node="crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.782920 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.805113 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.820300 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: W1124 17:26:48.824462 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-020a25dcd9a088a1530aeb9c0a14ce0ac49d952893240861b5eba0982cc07d86 WatchSource:0}: Error finding container 020a25dcd9a088a1530aeb9c0a14ce0ac49d952893240861b5eba0982cc07d86: Status 404 returned error can't find the container with id 020a25dcd9a088a1530aeb9c0a14ce0ac49d952893240861b5eba0982cc07d86 Nov 24 17:26:48 crc kubenswrapper[4808]: W1124 17:26:48.843411 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-9968cdd717cde0ec019644c51924b87c8ac7a9f29c5d093de814c5fadd03f506 WatchSource:0}: Error finding container 9968cdd717cde0ec019644c51924b87c8ac7a9f29c5d093de814c5fadd03f506: Status 404 returned error can't find the container with id 9968cdd717cde0ec019644c51924b87c8ac7a9f29c5d093de814c5fadd03f506 Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.844537 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: I1124 17:26:48.852467 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:26:48 crc kubenswrapper[4808]: W1124 17:26:48.883149 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-cc3ff23cdbba9ac8d6f0c90d0fe9a94728e76fcfb27fc51b8a55d25fcf3367b8 WatchSource:0}: Error finding container cc3ff23cdbba9ac8d6f0c90d0fe9a94728e76fcfb27fc51b8a55d25fcf3367b8: Status 404 returned error can't find the container with id cc3ff23cdbba9ac8d6f0c90d0fe9a94728e76fcfb27fc51b8a55d25fcf3367b8 Nov 24 17:26:48 crc kubenswrapper[4808]: E1124 17:26:48.896454 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.171:6443: connect: connection refused" interval="800ms" Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.113343 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.114654 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.114695 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.114706 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.114734 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 17:26:49 crc kubenswrapper[4808]: E1124 17:26:49.115236 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.171:6443: connect: connection refused" node="crc" Nov 24 17:26:49 crc kubenswrapper[4808]: W1124 17:26:49.175937 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.171:6443: connect: connection refused Nov 24 17:26:49 crc kubenswrapper[4808]: E1124 17:26:49.176049 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.171:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.287028 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 22:30:04.080976283 +0000 UTC Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.287873 4808 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.171:6443: connect: connection refused Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.353000 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"020a25dcd9a088a1530aeb9c0a14ce0ac49d952893240861b5eba0982cc07d86"} Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.354888 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cc3ff23cdbba9ac8d6f0c90d0fe9a94728e76fcfb27fc51b8a55d25fcf3367b8"} Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.356120 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"008bc696a93ff6e8893ea742d1d4a91cfe767a73c30f4be149749a8f6ce5fd61"} Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.357041 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"9968cdd717cde0ec019644c51924b87c8ac7a9f29c5d093de814c5fadd03f506"} Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.358276 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8b79126e6eb206ef4678763b08a25e44bfc7d9cf89c137101d845f700d371b96"} Nov 24 17:26:49 crc kubenswrapper[4808]: W1124 17:26:49.619899 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.171:6443: connect: connection refused Nov 24 17:26:49 crc kubenswrapper[4808]: E1124 17:26:49.620535 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.171:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:26:49 crc kubenswrapper[4808]: W1124 17:26:49.632172 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.171:6443: connect: connection refused Nov 24 17:26:49 crc kubenswrapper[4808]: E1124 17:26:49.632246 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.171:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:26:49 crc kubenswrapper[4808]: E1124 17:26:49.697845 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.171:6443: connect: connection refused" interval="1.6s" Nov 24 17:26:49 crc kubenswrapper[4808]: W1124 17:26:49.710068 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.171:6443: connect: connection refused Nov 24 17:26:49 crc kubenswrapper[4808]: E1124 17:26:49.710200 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.171:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.916046 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.918104 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.918140 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.918152 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:49 crc kubenswrapper[4808]: I1124 17:26:49.918182 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 17:26:49 crc kubenswrapper[4808]: E1124 17:26:49.918718 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.171:6443: connect: connection refused" node="crc" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.287530 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 07:05:20.726152598 +0000 UTC Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.287608 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1021h38m30.43854791s for next certificate rotation Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.288470 4808 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.171:6443: connect: connection refused Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.362989 4808 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6" exitCode=0 Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.363075 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6"} Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.363232 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.364432 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.364479 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.364492 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.365694 4808 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859" exitCode=0 Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.365786 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859"} Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.365825 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.366693 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.367623 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.367651 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.367659 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.368150 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.368204 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.368258 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.369664 4808 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="b1484dcf3f52f8d8388542b94e030ddab6f4fa2921d97269650d616ac488bca3" exitCode=0 Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.369754 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.369774 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"b1484dcf3f52f8d8388542b94e030ddab6f4fa2921d97269650d616ac488bca3"} Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.370772 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.370991 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.371085 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.371733 4808 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11" exitCode=0 Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.371772 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11"} Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.371810 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.373349 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.373375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.373383 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.375219 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6"} Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.375636 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438"} Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.375654 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214"} Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.375668 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b"} Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.375709 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.377945 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.378009 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:50 crc kubenswrapper[4808]: I1124 17:26:50.378046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:50 crc kubenswrapper[4808]: W1124 17:26:50.873363 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.171:6443: connect: connection refused Nov 24 17:26:50 crc kubenswrapper[4808]: E1124 17:26:50.873450 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.171:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.288734 4808 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.171:6443: connect: connection refused Nov 24 17:26:51 crc kubenswrapper[4808]: E1124 17:26:51.299806 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.171:6443: connect: connection refused" interval="3.2s" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.380720 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.381223 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3cfcd462f857faabd6f524d8ac5037283e272ba0705749a8df1a058a1279e753"} Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.382655 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.382706 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.382719 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.385896 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bdd30bb5725168f10a8a593c13456a101218f0bc17c492b7100470dee10fcf33"} Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.385940 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"23297e0c2ec1bd342e97252d23b64dcf6039bcdb46579119053d8c654f953a3b"} Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.385953 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3462fe4e148a5c80763c265a819dcb04af31902463e9f1e34afefce17b13bbb2"} Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.386069 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.386726 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.386751 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.386780 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.390909 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b"} Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.390968 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a"} Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.390984 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538"} Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.391000 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f"} Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.392875 4808 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816" exitCode=0 Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.393010 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.393629 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.393944 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816"} Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.395192 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.395229 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.395239 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.395506 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.395575 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.395585 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.519605 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.521124 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.521228 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.521249 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:51 crc kubenswrapper[4808]: I1124 17:26:51.521288 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 17:26:51 crc kubenswrapper[4808]: E1124 17:26:51.521799 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.171:6443: connect: connection refused" node="crc" Nov 24 17:26:51 crc kubenswrapper[4808]: W1124 17:26:51.987315 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.171:6443: connect: connection refused Nov 24 17:26:51 crc kubenswrapper[4808]: E1124 17:26:51.987402 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.171:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.398435 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5e0d18f4230d3ec037e3339cd2f725d060f528a4bd013a28c0350d25e3715586"} Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.398508 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.399297 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.399326 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.399339 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.401817 4808 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6" exitCode=0 Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.401911 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.401937 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.402495 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6"} Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.402557 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.402560 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.403420 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.403436 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.403454 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.403465 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.403424 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.403494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.403504 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.403538 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:52 crc kubenswrapper[4808]: I1124 17:26:52.403568 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.137529 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.385253 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.412499 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c"} Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.412549 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.412589 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9"} Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.412624 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89"} Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.412646 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82"} Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.412666 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4"} Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.412532 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.412690 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.412776 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.414806 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.414846 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.414849 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.414889 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.414906 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.414857 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.415053 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.415070 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.415079 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.599003 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.599192 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.600412 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.600442 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:53 crc kubenswrapper[4808]: I1124 17:26:53.600452 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:54 crc kubenswrapper[4808]: I1124 17:26:54.112128 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 24 17:26:54 crc kubenswrapper[4808]: I1124 17:26:54.416244 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:54 crc kubenswrapper[4808]: I1124 17:26:54.416252 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 17:26:54 crc kubenswrapper[4808]: I1124 17:26:54.416481 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:54 crc kubenswrapper[4808]: I1124 17:26:54.417806 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:54 crc kubenswrapper[4808]: I1124 17:26:54.417929 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:54 crc kubenswrapper[4808]: I1124 17:26:54.417941 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:54 crc kubenswrapper[4808]: I1124 17:26:54.417994 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:54 crc kubenswrapper[4808]: I1124 17:26:54.417952 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:54 crc kubenswrapper[4808]: I1124 17:26:54.418067 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:54 crc kubenswrapper[4808]: I1124 17:26:54.722600 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:54 crc kubenswrapper[4808]: I1124 17:26:54.724466 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:54 crc kubenswrapper[4808]: I1124 17:26:54.724510 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:54 crc kubenswrapper[4808]: I1124 17:26:54.724523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:54 crc kubenswrapper[4808]: I1124 17:26:54.724557 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 17:26:55 crc kubenswrapper[4808]: I1124 17:26:55.419596 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:55 crc kubenswrapper[4808]: I1124 17:26:55.420856 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:55 crc kubenswrapper[4808]: I1124 17:26:55.420905 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:55 crc kubenswrapper[4808]: I1124 17:26:55.420923 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:56 crc kubenswrapper[4808]: I1124 17:26:56.599889 4808 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 17:26:56 crc kubenswrapper[4808]: I1124 17:26:56.600072 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 17:26:57 crc kubenswrapper[4808]: I1124 17:26:57.342511 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:26:57 crc kubenswrapper[4808]: I1124 17:26:57.342708 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:57 crc kubenswrapper[4808]: I1124 17:26:57.343853 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:57 crc kubenswrapper[4808]: I1124 17:26:57.343881 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:57 crc kubenswrapper[4808]: I1124 17:26:57.343891 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:57 crc kubenswrapper[4808]: I1124 17:26:57.371146 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:26:57 crc kubenswrapper[4808]: I1124 17:26:57.371354 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:57 crc kubenswrapper[4808]: I1124 17:26:57.372554 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:57 crc kubenswrapper[4808]: I1124 17:26:57.372586 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:57 crc kubenswrapper[4808]: I1124 17:26:57.372596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:58 crc kubenswrapper[4808]: E1124 17:26:58.418400 4808 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 17:26:58 crc kubenswrapper[4808]: I1124 17:26:58.977150 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:26:58 crc kubenswrapper[4808]: I1124 17:26:58.977296 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:58 crc kubenswrapper[4808]: I1124 17:26:58.978531 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:58 crc kubenswrapper[4808]: I1124 17:26:58.978586 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:58 crc kubenswrapper[4808]: I1124 17:26:58.978596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:26:58 crc kubenswrapper[4808]: I1124 17:26:58.982325 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:26:59 crc kubenswrapper[4808]: I1124 17:26:59.141780 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:26:59 crc kubenswrapper[4808]: I1124 17:26:59.145628 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:26:59 crc kubenswrapper[4808]: I1124 17:26:59.429248 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:26:59 crc kubenswrapper[4808]: I1124 17:26:59.430510 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:26:59 crc kubenswrapper[4808]: I1124 17:26:59.430581 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:26:59 crc kubenswrapper[4808]: I1124 17:26:59.430596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:00 crc kubenswrapper[4808]: I1124 17:27:00.432407 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:27:00 crc kubenswrapper[4808]: I1124 17:27:00.433984 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:00 crc kubenswrapper[4808]: I1124 17:27:00.434066 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:00 crc kubenswrapper[4808]: I1124 17:27:00.434086 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:01 crc kubenswrapper[4808]: I1124 17:27:01.895222 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 24 17:27:01 crc kubenswrapper[4808]: I1124 17:27:01.895579 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:27:01 crc kubenswrapper[4808]: I1124 17:27:01.897415 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:01 crc kubenswrapper[4808]: I1124 17:27:01.897470 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:01 crc kubenswrapper[4808]: I1124 17:27:01.897482 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:02 crc kubenswrapper[4808]: I1124 17:27:02.291474 4808 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 24 17:27:02 crc kubenswrapper[4808]: W1124 17:27:02.323788 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 24 17:27:02 crc kubenswrapper[4808]: I1124 17:27:02.323882 4808 trace.go:236] Trace[655343791]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 17:26:52.322) (total time: 10001ms): Nov 24 17:27:02 crc kubenswrapper[4808]: Trace[655343791]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (17:27:02.323) Nov 24 17:27:02 crc kubenswrapper[4808]: Trace[655343791]: [10.001088577s] [10.001088577s] END Nov 24 17:27:02 crc kubenswrapper[4808]: E1124 17:27:02.323907 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 24 17:27:02 crc kubenswrapper[4808]: I1124 17:27:02.438353 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 17:27:02 crc kubenswrapper[4808]: I1124 17:27:02.440205 4808 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5e0d18f4230d3ec037e3339cd2f725d060f528a4bd013a28c0350d25e3715586" exitCode=255 Nov 24 17:27:02 crc kubenswrapper[4808]: I1124 17:27:02.440267 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5e0d18f4230d3ec037e3339cd2f725d060f528a4bd013a28c0350d25e3715586"} Nov 24 17:27:02 crc kubenswrapper[4808]: I1124 17:27:02.440461 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:27:02 crc kubenswrapper[4808]: I1124 17:27:02.441566 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:02 crc kubenswrapper[4808]: I1124 17:27:02.441607 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:02 crc kubenswrapper[4808]: I1124 17:27:02.441620 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:02 crc kubenswrapper[4808]: I1124 17:27:02.442227 4808 scope.go:117] "RemoveContainer" containerID="5e0d18f4230d3ec037e3339cd2f725d060f528a4bd013a28c0350d25e3715586" Nov 24 17:27:02 crc kubenswrapper[4808]: I1124 17:27:02.826293 4808 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 17:27:02 crc kubenswrapper[4808]: I1124 17:27:02.826724 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 17:27:02 crc kubenswrapper[4808]: I1124 17:27:02.833944 4808 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 17:27:02 crc kubenswrapper[4808]: I1124 17:27:02.834143 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 17:27:03 crc kubenswrapper[4808]: I1124 17:27:03.157810 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:27:03 crc kubenswrapper[4808]: I1124 17:27:03.391876 4808 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]log ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]etcd ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/openshift.io-startkubeinformers ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/openshift.io-api-request-count-filter ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/generic-apiserver-start-informers ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/priority-and-fairness-config-consumer ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/priority-and-fairness-filter ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/start-apiextensions-informers ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/start-apiextensions-controllers ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/crd-informer-synced ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/start-system-namespaces-controller ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/start-cluster-authentication-info-controller ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/start-legacy-token-tracking-controller ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/start-service-ip-repair-controllers ok Nov 24 17:27:03 crc kubenswrapper[4808]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Nov 24 17:27:03 crc kubenswrapper[4808]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/priority-and-fairness-config-producer ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/bootstrap-controller ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/start-kube-aggregator-informers ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/apiservice-status-local-available-controller ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/apiservice-status-remote-available-controller ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/apiservice-registration-controller ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/apiservice-wait-for-first-sync ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/apiservice-discovery-controller ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/kube-apiserver-autoregistration ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]autoregister-completion ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/apiservice-openapi-controller ok Nov 24 17:27:03 crc kubenswrapper[4808]: [+]poststarthook/apiservice-openapiv3-controller ok Nov 24 17:27:03 crc kubenswrapper[4808]: livez check failed Nov 24 17:27:03 crc kubenswrapper[4808]: I1124 17:27:03.391961 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:27:03 crc kubenswrapper[4808]: I1124 17:27:03.446236 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 17:27:03 crc kubenswrapper[4808]: I1124 17:27:03.448654 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb"} Nov 24 17:27:03 crc kubenswrapper[4808]: I1124 17:27:03.448808 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:27:03 crc kubenswrapper[4808]: I1124 17:27:03.450093 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:03 crc kubenswrapper[4808]: I1124 17:27:03.450139 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:03 crc kubenswrapper[4808]: I1124 17:27:03.450151 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:04 crc kubenswrapper[4808]: I1124 17:27:04.451491 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:27:04 crc kubenswrapper[4808]: I1124 17:27:04.451581 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:27:04 crc kubenswrapper[4808]: I1124 17:27:04.452304 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:04 crc kubenswrapper[4808]: I1124 17:27:04.452336 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:04 crc kubenswrapper[4808]: I1124 17:27:04.452346 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:05 crc kubenswrapper[4808]: I1124 17:27:05.453362 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:27:05 crc kubenswrapper[4808]: I1124 17:27:05.454963 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:05 crc kubenswrapper[4808]: I1124 17:27:05.455038 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:05 crc kubenswrapper[4808]: I1124 17:27:05.455054 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:06 crc kubenswrapper[4808]: I1124 17:27:06.182769 4808 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 24 17:27:06 crc kubenswrapper[4808]: I1124 17:27:06.599980 4808 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 17:27:06 crc kubenswrapper[4808]: I1124 17:27:06.600072 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 17:27:07 crc kubenswrapper[4808]: E1124 17:27:07.823930 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 24 17:27:07 crc kubenswrapper[4808]: I1124 17:27:07.826886 4808 trace.go:236] Trace[578269031]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 17:26:57.811) (total time: 10015ms): Nov 24 17:27:07 crc kubenswrapper[4808]: Trace[578269031]: ---"Objects listed" error: 10015ms (17:27:07.826) Nov 24 17:27:07 crc kubenswrapper[4808]: Trace[578269031]: [10.015520162s] [10.015520162s] END Nov 24 17:27:07 crc kubenswrapper[4808]: I1124 17:27:07.826960 4808 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 24 17:27:07 crc kubenswrapper[4808]: I1124 17:27:07.827489 4808 trace.go:236] Trace[1015688320]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 17:26:54.337) (total time: 13490ms): Nov 24 17:27:07 crc kubenswrapper[4808]: Trace[1015688320]: ---"Objects listed" error: 13490ms (17:27:07.827) Nov 24 17:27:07 crc kubenswrapper[4808]: Trace[1015688320]: [13.490195843s] [13.490195843s] END Nov 24 17:27:07 crc kubenswrapper[4808]: I1124 17:27:07.827510 4808 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 24 17:27:07 crc kubenswrapper[4808]: I1124 17:27:07.827659 4808 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 24 17:27:07 crc kubenswrapper[4808]: E1124 17:27:07.827726 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 24 17:27:07 crc kubenswrapper[4808]: I1124 17:27:07.829158 4808 trace.go:236] Trace[1746084301]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 17:26:52.858) (total time: 14970ms): Nov 24 17:27:07 crc kubenswrapper[4808]: Trace[1746084301]: ---"Objects listed" error: 14970ms (17:27:07.828) Nov 24 17:27:07 crc kubenswrapper[4808]: Trace[1746084301]: [14.970565868s] [14.970565868s] END Nov 24 17:27:07 crc kubenswrapper[4808]: I1124 17:27:07.829184 4808 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.279782 4808 apiserver.go:52] "Watching apiserver" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.282732 4808 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.282999 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-k9pwr","openshift-image-registry/node-ca-7cbtg"] Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.283426 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.283454 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.283810 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.284063 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.284066 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.284186 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.284280 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.284600 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.284650 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7cbtg" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.284691 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.284824 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-k9pwr" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.288648 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.288717 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.288826 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.288863 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.288917 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.288927 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.288933 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.288994 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.289127 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.289132 4808 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.289240 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.289296 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.289322 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.289552 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.289753 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.289844 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.289928 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.306503 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.318853 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330529 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330604 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330636 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330664 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330692 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330718 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330747 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330776 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330807 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330834 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330864 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330888 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330918 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330943 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330967 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.330990 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331033 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331058 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331081 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331104 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331127 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331152 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331174 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331202 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331228 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331257 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331284 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331242 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331314 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331345 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331374 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331441 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331454 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331499 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331486 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331599 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331639 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331659 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331689 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331715 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331737 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331766 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331801 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331830 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331852 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331880 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331899 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331909 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331891 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331939 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331972 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.331998 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332004 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332038 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332066 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332061 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332092 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332118 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332142 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332166 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332193 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332216 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332226 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332241 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332270 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332277 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332291 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332315 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332335 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332353 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332373 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332358 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332392 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332413 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332437 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332461 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332488 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332514 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332535 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332558 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332583 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332603 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332634 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332760 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332790 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332808 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332825 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332842 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332860 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332877 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332898 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332915 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332931 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332956 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332976 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332994 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333048 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333074 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333114 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333135 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333157 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333177 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333196 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333215 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333234 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333254 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333271 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333290 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333307 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333327 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333347 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333367 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333387 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333405 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333423 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333440 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333457 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333475 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333494 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333511 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333530 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333549 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333596 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333616 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333639 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333659 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333677 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333693 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333712 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333747 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333764 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333783 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333803 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333822 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333841 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333858 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333876 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333892 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333910 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333926 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333943 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333963 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333996 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334027 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334049 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334071 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334094 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334117 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334138 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334156 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334172 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334189 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334206 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334225 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334244 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334264 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334282 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334303 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334325 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334343 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334361 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334378 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334398 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334416 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334432 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334449 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334468 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334485 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334501 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334518 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334534 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334549 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334566 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334584 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334617 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334646 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334666 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334686 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334705 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334723 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334740 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334762 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334807 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334829 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334848 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334877 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334903 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334925 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334941 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334965 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.334989 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.335008 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336267 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336302 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336328 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336349 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336370 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336390 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336409 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336430 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336450 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336471 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336490 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336509 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336528 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336547 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336698 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336738 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336759 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336781 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336802 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336824 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336857 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ded1d94f-dc75-45de-b921-a66747cd3f18-host\") pod \"node-ca-7cbtg\" (UID: \"ded1d94f-dc75-45de-b921-a66747cd3f18\") " pod="openshift-image-registry/node-ca-7cbtg" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336886 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c07994f6-627f-4641-b164-ae81aac54782-hosts-file\") pod \"node-resolver-k9pwr\" (UID: \"c07994f6-627f-4641-b164-ae81aac54782\") " pod="openshift-dns/node-resolver-k9pwr" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336907 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336928 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7wjj\" (UniqueName: \"kubernetes.io/projected/ded1d94f-dc75-45de-b921-a66747cd3f18-kube-api-access-p7wjj\") pod \"node-ca-7cbtg\" (UID: \"ded1d94f-dc75-45de-b921-a66747cd3f18\") " pod="openshift-image-registry/node-ca-7cbtg" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336952 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.336977 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr4gx\" (UniqueName: \"kubernetes.io/projected/c07994f6-627f-4641-b164-ae81aac54782-kube-api-access-qr4gx\") pod \"node-resolver-k9pwr\" (UID: \"c07994f6-627f-4641-b164-ae81aac54782\") " pod="openshift-dns/node-resolver-k9pwr" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337001 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337038 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337059 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337094 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337119 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337139 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ded1d94f-dc75-45de-b921-a66747cd3f18-serviceca\") pod \"node-ca-7cbtg\" (UID: \"ded1d94f-dc75-45de-b921-a66747cd3f18\") " pod="openshift-image-registry/node-ca-7cbtg" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337160 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337243 4808 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337256 4808 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337267 4808 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337355 4808 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337369 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337380 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337404 4808 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337415 4808 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337426 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337437 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337448 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.338146 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332602 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.340782 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332650 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332712 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332795 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.332977 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.333554 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.335174 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.335235 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.335706 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.335889 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337062 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337410 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337399 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.337626 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.338145 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.338228 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.338324 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.338629 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.338698 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.338707 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.338791 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.338817 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.338903 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.339640 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.339704 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.340091 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.340159 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.340257 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.340410 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.340552 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.340711 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.341504 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.341677 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:27:08.841645975 +0000 UTC m=+21.439313847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.341987 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.342156 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.342358 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.342417 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.342438 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.342814 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.343141 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.343162 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.343164 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.344239 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.344243 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.344287 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.344294 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.344741 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.344920 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.345389 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.346224 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.346754 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.347081 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.347127 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.347156 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.347296 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.347624 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.347641 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.347661 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.347879 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.347948 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.347996 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.348001 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.348335 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.348527 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.348686 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.348841 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.349148 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.349276 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.349402 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.349625 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.349696 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.349741 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.350046 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.350045 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.350891 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.350899 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.351280 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.351279 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.351399 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.351555 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.351652 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.351703 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.351760 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.351996 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.352009 4808 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.352096 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.352293 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.352417 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.352746 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.353039 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.353369 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.353540 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.353580 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.353683 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.353979 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.354235 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.354268 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.354778 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.354853 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:08.854829134 +0000 UTC m=+21.452497136 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.355090 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.355104 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.355215 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.355910 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.356082 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.356326 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.356684 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.357570 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.357878 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.358363 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.358446 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.358486 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.358488 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.359266 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.359349 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:08.859325133 +0000 UTC m=+21.456992935 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.360633 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.361857 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.363275 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.363466 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.363690 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.363802 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.364368 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.364425 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.364818 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.365050 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.365100 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.374077 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.374114 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.374131 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.374212 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:08.87418319 +0000 UTC m=+21.471851192 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.383191 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.383230 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.383243 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.383305 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:08.883283452 +0000 UTC m=+21.480951254 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.387805 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.394137 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.397919 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.400306 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.401090 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.401152 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.401804 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.402389 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.404111 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.405835 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.406073 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.406526 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.408459 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.408581 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.409063 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.409916 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.410490 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.410574 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.413921 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.414398 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.415224 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.415257 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.415405 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.418902 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.419991 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.419990 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.420670 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.421808 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.423505 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.423743 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.423893 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.424703 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.426114 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.426506 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.428251 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.429290 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.430133 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.430487 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.430587 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.431194 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.431627 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.431803 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.432080 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.432140 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.432402 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.432533 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.433449 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.436614 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438028 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ded1d94f-dc75-45de-b921-a66747cd3f18-serviceca\") pod \"node-ca-7cbtg\" (UID: \"ded1d94f-dc75-45de-b921-a66747cd3f18\") " pod="openshift-image-registry/node-ca-7cbtg" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438058 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438104 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c07994f6-627f-4641-b164-ae81aac54782-hosts-file\") pod \"node-resolver-k9pwr\" (UID: \"c07994f6-627f-4641-b164-ae81aac54782\") " pod="openshift-dns/node-resolver-k9pwr" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438122 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ded1d94f-dc75-45de-b921-a66747cd3f18-host\") pod \"node-ca-7cbtg\" (UID: \"ded1d94f-dc75-45de-b921-a66747cd3f18\") " pod="openshift-image-registry/node-ca-7cbtg" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438159 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7wjj\" (UniqueName: \"kubernetes.io/projected/ded1d94f-dc75-45de-b921-a66747cd3f18-kube-api-access-p7wjj\") pod \"node-ca-7cbtg\" (UID: \"ded1d94f-dc75-45de-b921-a66747cd3f18\") " pod="openshift-image-registry/node-ca-7cbtg" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438177 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438194 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr4gx\" (UniqueName: \"kubernetes.io/projected/c07994f6-627f-4641-b164-ae81aac54782-kube-api-access-qr4gx\") pod \"node-resolver-k9pwr\" (UID: \"c07994f6-627f-4641-b164-ae81aac54782\") " pod="openshift-dns/node-resolver-k9pwr" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438263 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438274 4808 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438285 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438293 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438303 4808 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438312 4808 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438321 4808 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438330 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438339 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438347 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438355 4808 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438370 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438379 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438387 4808 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438396 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438405 4808 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438413 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438422 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438431 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438440 4808 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438449 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438458 4808 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438467 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438480 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438489 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438497 4808 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438507 4808 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438516 4808 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438525 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438534 4808 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438543 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438552 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438560 4808 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438570 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438579 4808 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438588 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438597 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438606 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438614 4808 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438622 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438631 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438639 4808 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438648 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438656 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438665 4808 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438673 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438682 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438690 4808 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438699 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438708 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438717 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438725 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438734 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438743 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438751 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438761 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438769 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438778 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438786 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438795 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438803 4808 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438812 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438820 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438829 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438850 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438859 4808 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438868 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438877 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438886 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438894 4808 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438904 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438912 4808 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438921 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438931 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438940 4808 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438948 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438957 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438966 4808 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438975 4808 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438983 4808 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.438991 4808 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439000 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439009 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439029 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439044 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439053 4808 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439062 4808 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439072 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439081 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439090 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439100 4808 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439108 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439117 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439125 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439134 4808 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439142 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439151 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439159 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439168 4808 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439176 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439184 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439192 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439202 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439213 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439230 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439243 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439255 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439254 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439268 4808 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439664 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439674 4808 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439683 4808 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439691 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439700 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439710 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439719 4808 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439728 4808 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439736 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439744 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439752 4808 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439768 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439776 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439784 4808 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439792 4808 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439801 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439809 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439818 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439825 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439834 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439842 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439850 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439858 4808 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439866 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439874 4808 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439883 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439891 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439899 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439907 4808 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439916 4808 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439926 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439934 4808 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439945 4808 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439953 4808 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439960 4808 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439968 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439977 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439985 4808 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.439993 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.440002 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.440011 4808 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.440059 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.440067 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.440211 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ded1d94f-dc75-45de-b921-a66747cd3f18-host\") pod \"node-ca-7cbtg\" (UID: \"ded1d94f-dc75-45de-b921-a66747cd3f18\") " pod="openshift-image-registry/node-ca-7cbtg" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.440243 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.440533 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ded1d94f-dc75-45de-b921-a66747cd3f18-serviceca\") pod \"node-ca-7cbtg\" (UID: \"ded1d94f-dc75-45de-b921-a66747cd3f18\") " pod="openshift-image-registry/node-ca-7cbtg" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.440557 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c07994f6-627f-4641-b164-ae81aac54782-hosts-file\") pod \"node-resolver-k9pwr\" (UID: \"c07994f6-627f-4641-b164-ae81aac54782\") " pod="openshift-dns/node-resolver-k9pwr" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.440724 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.442214 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.442318 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.442560 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.443368 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.443009 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.443739 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.444602 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.445594 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.446615 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.446702 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.449037 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.450498 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.453133 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.456374 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.456477 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.456538 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.457499 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.457917 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.458269 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.458564 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.458672 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.459028 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.459333 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.459389 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.459431 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.459882 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.459937 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.463176 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr4gx\" (UniqueName: \"kubernetes.io/projected/c07994f6-627f-4641-b164-ae81aac54782-kube-api-access-qr4gx\") pod \"node-resolver-k9pwr\" (UID: \"c07994f6-627f-4641-b164-ae81aac54782\") " pod="openshift-dns/node-resolver-k9pwr" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.464585 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7wjj\" (UniqueName: \"kubernetes.io/projected/ded1d94f-dc75-45de-b921-a66747cd3f18-kube-api-access-p7wjj\") pod \"node-ca-7cbtg\" (UID: \"ded1d94f-dc75-45de-b921-a66747cd3f18\") " pod="openshift-image-registry/node-ca-7cbtg" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.469154 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.469748 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.469755 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.470102 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.470149 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.472354 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.481717 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.483928 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.484633 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.486631 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.487263 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.487296 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.488531 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.489165 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.490984 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.491526 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.492047 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.493318 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.493972 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.495042 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.497985 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.498662 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.499170 4808 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.499293 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.500291 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.501788 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.502301 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.504797 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.506802 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.507376 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.507654 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.508575 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.509310 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.510410 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.510888 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.511868 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.512512 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.513536 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.514130 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.515047 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.515405 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.515582 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.516659 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.517131 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.517956 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.518550 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.519090 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.520030 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.520491 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.521513 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.524760 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.526263 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.533929 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540769 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540800 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540815 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540826 4808 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540837 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540847 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540858 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540868 4808 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540878 4808 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540888 4808 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540900 4808 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540912 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540925 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540937 4808 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540950 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540962 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540974 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540986 4808 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.540998 4808 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.541009 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.541041 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.541053 4808 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.541064 4808 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.541077 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.541088 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.541100 4808 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.541110 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.541123 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.541134 4808 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.541145 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.541157 4808 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.541170 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.542878 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.555482 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.566184 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.579709 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.599048 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.599427 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.611162 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.614808 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.621103 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.632313 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.634167 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7cbtg" Nov 24 17:27:08 crc kubenswrapper[4808]: W1124 17:27:08.637467 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-675dbcd451a62be55a98b896f7263706148b09c294dbd88aa204d04296a08fdd WatchSource:0}: Error finding container 675dbcd451a62be55a98b896f7263706148b09c294dbd88aa204d04296a08fdd: Status 404 returned error can't find the container with id 675dbcd451a62be55a98b896f7263706148b09c294dbd88aa204d04296a08fdd Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.645525 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.649141 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-k9pwr" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.660416 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.672947 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.680970 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.688473 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: W1124 17:27:08.690488 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podded1d94f_dc75_45de_b921_a66747cd3f18.slice/crio-bd703f819dbe5dc21049b6f9c631a0f0391fab509cc2f03c51f4e2f6c01a3a62 WatchSource:0}: Error finding container bd703f819dbe5dc21049b6f9c631a0f0391fab509cc2f03c51f4e2f6c01a3a62: Status 404 returned error can't find the container with id bd703f819dbe5dc21049b6f9c631a0f0391fab509cc2f03c51f4e2f6c01a3a62 Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.714828 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.762539 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-zg5rq"] Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.764469 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-q5m49"] Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.764731 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.764908 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5b2z8"] Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.765097 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.766190 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.773000 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-ht87w"] Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.772998 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.773534 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.773563 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.773582 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.773108 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.773115 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.773207 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.773255 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.773293 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.773312 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.773378 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.773396 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.773446 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.773449 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.773464 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.788845 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.790112 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.790551 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.790734 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.790991 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.797409 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.811385 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.835307 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844430 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844600 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-multus-socket-dir-parent\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844639 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-var-lib-cni-multus\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844663 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6a192cd6-853d-4643-855d-ac61168b1cda-multus-daemon-config\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844687 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-env-overrides\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844709 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-var-lib-kubelet\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844735 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-run-multus-certs\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844759 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-etc-kubernetes\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844780 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-ovnkube-config\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844808 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zgw2\" (UniqueName: \"kubernetes.io/projected/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-kube-api-access-5zgw2\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844852 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-hostroot\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844877 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-var-lib-cni-bin\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844901 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6tl6\" (UniqueName: \"kubernetes.io/projected/6a192cd6-853d-4643-855d-ac61168b1cda-kube-api-access-p6tl6\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844923 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-systemd-units\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844948 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-run-netns\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844968 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhss2\" (UniqueName: \"kubernetes.io/projected/5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0-kube-api-access-dhss2\") pod \"machine-config-daemon-ht87w\" (UID: \"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\") " pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.844984 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-log-socket\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845004 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-cni-binary-copy\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845035 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0-rootfs\") pod \"machine-config-daemon-ht87w\" (UID: \"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\") " pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845052 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-kubelet\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845067 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-run-netns\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845082 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-ovn\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845097 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-etc-openvswitch\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845117 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrc9l\" (UniqueName: \"kubernetes.io/projected/4caad6d6-912a-4441-aa90-38a0e29591f5-kube-api-access-lrc9l\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845135 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-os-release\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845161 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0-proxy-tls\") pod \"machine-config-daemon-ht87w\" (UID: \"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\") " pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845192 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-cnibin\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845217 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-run-k8s-cni-cncf-io\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845240 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-cni-bin\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845259 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-os-release\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845294 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-cni-netd\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845320 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-ovnkube-script-lib\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845345 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845368 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-system-cni-dir\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845386 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-var-lib-openvswitch\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845400 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6a192cd6-853d-4643-855d-ac61168b1cda-cni-binary-copy\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845415 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-openvswitch\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845431 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-system-cni-dir\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845453 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-multus-cni-dir\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845468 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-multus-conf-dir\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845485 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845499 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-slash\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845515 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-node-log\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845530 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-cnibin\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845559 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0-mcd-auth-proxy-config\") pod \"machine-config-daemon-ht87w\" (UID: \"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\") " pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845577 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845592 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-run-ovn-kubernetes\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845610 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4caad6d6-912a-4441-aa90-38a0e29591f5-ovn-node-metrics-cert\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.845629 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-systemd\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.845746 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:27:09.845725271 +0000 UTC m=+22.443393073 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.866565 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.889607 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.908379 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.937352 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946194 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-cnibin\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946248 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-os-release\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946291 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0-proxy-tls\") pod \"machine-config-daemon-ht87w\" (UID: \"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\") " pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946319 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946340 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-run-k8s-cni-cncf-io\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946362 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-cni-bin\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946381 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-os-release\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946375 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-cnibin\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946488 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-system-cni-dir\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946401 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-system-cni-dir\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946557 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-var-lib-openvswitch\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946581 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-cni-netd\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946616 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-ovnkube-script-lib\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946639 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946685 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-system-cni-dir\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946705 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6a192cd6-853d-4643-855d-ac61168b1cda-cni-binary-copy\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946706 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-os-release\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946726 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-openvswitch\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946774 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-multus-cni-dir\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.946858 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-openvswitch\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.947063 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-os-release\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.947423 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.947462 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-multus-cni-dir\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.947589 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:09.947531277 +0000 UTC m=+22.545199079 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.947721 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-cni-bin\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.947789 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-system-cni-dir\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.947835 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-cni-netd\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.947860 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-var-lib-openvswitch\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.947879 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-ovnkube-script-lib\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.947891 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-run-k8s-cni-cncf-io\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.947925 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-multus-conf-dir\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.947970 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-multus-conf-dir\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.947996 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948080 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948132 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948150 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948188 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948214 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0-mcd-auth-proxy-config\") pod \"machine-config-daemon-ht87w\" (UID: \"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\") " pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948238 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-slash\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.948260 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948267 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-node-log\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.948282 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.948299 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948310 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-cnibin\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.948355 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:09.94833373 +0000 UTC m=+22.546001532 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948381 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948391 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-cnibin\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948411 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-systemd\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948435 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-run-ovn-kubernetes\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948452 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-slash\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948455 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4caad6d6-912a-4441-aa90-38a0e29591f5-ovn-node-metrics-cert\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948489 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6a192cd6-853d-4643-855d-ac61168b1cda-multus-daemon-config\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948508 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-env-overrides\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948528 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-multus-socket-dir-parent\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948546 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-var-lib-cni-multus\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948562 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-etc-kubernetes\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948579 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-ovnkube-config\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948595 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zgw2\" (UniqueName: \"kubernetes.io/projected/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-kube-api-access-5zgw2\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948612 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-var-lib-kubelet\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948632 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-run-multus-certs\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948649 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-hostroot\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948672 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948699 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-var-lib-cni-bin\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948719 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6tl6\" (UniqueName: \"kubernetes.io/projected/6a192cd6-853d-4643-855d-ac61168b1cda-kube-api-access-p6tl6\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948748 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-systemd-units\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948778 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhss2\" (UniqueName: \"kubernetes.io/projected/5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0-kube-api-access-dhss2\") pod \"machine-config-daemon-ht87w\" (UID: \"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\") " pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948802 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-log-socket\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948827 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-run-netns\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948848 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-cni-binary-copy\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948878 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-kubelet\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948911 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-run-netns\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948937 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-ovn\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948964 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0-rootfs\") pod \"machine-config-daemon-ht87w\" (UID: \"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\") " pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.948987 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-etc-openvswitch\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.949002 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0-mcd-auth-proxy-config\") pod \"machine-config-daemon-ht87w\" (UID: \"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\") " pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.949007 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrc9l\" (UniqueName: \"kubernetes.io/projected/4caad6d6-912a-4441-aa90-38a0e29591f5-kube-api-access-lrc9l\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.949141 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.949169 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-systemd\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.949211 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:09.949193844 +0000 UTC m=+22.546861646 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.949227 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-run-ovn-kubernetes\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.949261 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-etc-kubernetes\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.949289 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-node-log\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.949311 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.949327 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.949360 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:08 crc kubenswrapper[4808]: E1124 17:27:08.949394 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:09.94938552 +0000 UTC m=+22.547053542 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.949494 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-systemd-units\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.949644 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-log-socket\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.949668 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-run-netns\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.949826 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-run-multus-certs\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.950106 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-var-lib-kubelet\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.950106 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-ovnkube-config\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.950137 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-hostroot\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.949359 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-var-lib-cni-bin\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.950174 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-ovn\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.950213 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-multus-socket-dir-parent\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.950213 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-kubelet\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.950239 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-run-netns\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.950290 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6a192cd6-853d-4643-855d-ac61168b1cda-host-var-lib-cni-multus\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.950334 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0-rootfs\") pod \"machine-config-daemon-ht87w\" (UID: \"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\") " pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.950389 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-etc-openvswitch\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.950643 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-env-overrides\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.951635 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6a192cd6-853d-4643-855d-ac61168b1cda-cni-binary-copy\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.952868 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4caad6d6-912a-4441-aa90-38a0e29591f5-ovn-node-metrics-cert\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.953266 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0-proxy-tls\") pod \"machine-config-daemon-ht87w\" (UID: \"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\") " pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.957394 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.958636 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6a192cd6-853d-4643-855d-ac61168b1cda-multus-daemon-config\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.961320 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-cni-binary-copy\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.978922 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.991663 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6tl6\" (UniqueName: \"kubernetes.io/projected/6a192cd6-853d-4643-855d-ac61168b1cda-kube-api-access-p6tl6\") pod \"multus-q5m49\" (UID: \"6a192cd6-853d-4643-855d-ac61168b1cda\") " pod="openshift-multus/multus-q5m49" Nov 24 17:27:08 crc kubenswrapper[4808]: I1124 17:27:08.996918 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhss2\" (UniqueName: \"kubernetes.io/projected/5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0-kube-api-access-dhss2\") pod \"machine-config-daemon-ht87w\" (UID: \"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\") " pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:08.998532 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrc9l\" (UniqueName: \"kubernetes.io/projected/4caad6d6-912a-4441-aa90-38a0e29591f5-kube-api-access-lrc9l\") pod \"ovnkube-node-5b2z8\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.002362 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d18f4230d3ec037e3339cd2f725d060f528a4bd013a28c0350d25e3715586\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:02Z\\\",\\\"message\\\":\\\"W1124 17:26:51.512994 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 17:26:51.513507 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764005211 cert, and key in /tmp/serving-cert-2994416005/serving-signer.crt, /tmp/serving-cert-2994416005/serving-signer.key\\\\nI1124 17:26:51.887454 1 observer_polling.go:159] Starting file observer\\\\nW1124 17:26:51.890403 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 17:26:51.890599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:26:51.891348 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2994416005/tls.crt::/tmp/serving-cert-2994416005/tls.key\\\\\\\"\\\\nF1124 17:27:02.148785 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.003204 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zgw2\" (UniqueName: \"kubernetes.io/projected/fe424e29-3d8f-49e3-ac0c-33dc63f93cf8-kube-api-access-5zgw2\") pod \"multus-additional-cni-plugins-zg5rq\" (UID: \"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\") " pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.020472 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.033323 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.044751 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.058396 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.066483 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.076294 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.095542 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.115377 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d18f4230d3ec037e3339cd2f725d060f528a4bd013a28c0350d25e3715586\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:02Z\\\",\\\"message\\\":\\\"W1124 17:26:51.512994 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 17:26:51.513507 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764005211 cert, and key in /tmp/serving-cert-2994416005/serving-signer.crt, /tmp/serving-cert-2994416005/serving-signer.key\\\\nI1124 17:26:51.887454 1 observer_polling.go:159] Starting file observer\\\\nW1124 17:26:51.890403 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 17:26:51.890599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:26:51.891348 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2994416005/tls.crt::/tmp/serving-cert-2994416005/tls.key\\\\\\\"\\\\nF1124 17:27:02.148785 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.128350 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.138461 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-q5m49" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.139731 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.152311 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: W1124 17:27:09.154135 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a192cd6_853d_4643_855d_ac61168b1cda.slice/crio-31d35c20e79f73aa4124bc9bc8da0f3ab2925eeb43022667178b66083469446f WatchSource:0}: Error finding container 31d35c20e79f73aa4124bc9bc8da0f3ab2925eeb43022667178b66083469446f: Status 404 returned error can't find the container with id 31d35c20e79f73aa4124bc9bc8da0f3ab2925eeb43022667178b66083469446f Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.157493 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.167811 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.170361 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.183465 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: W1124 17:27:09.191262 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4caad6d6_912a_4441_aa90_38a0e29591f5.slice/crio-e2e107b74a075760d5543d3e2da94b5776dcdda42690920b33b3ca2d8febdac2 WatchSource:0}: Error finding container e2e107b74a075760d5543d3e2da94b5776dcdda42690920b33b3ca2d8febdac2: Status 404 returned error can't find the container with id e2e107b74a075760d5543d3e2da94b5776dcdda42690920b33b3ca2d8febdac2 Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.191464 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.209789 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: W1124 17:27:09.222944 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b86a0e3_4cb3_4ca4_a9ac_c66e137415d0.slice/crio-f1c14ae3fe21779f0b3c1ab21cb260b6a6732f8a6204b5aa0c90c5855bb36548 WatchSource:0}: Error finding container f1c14ae3fe21779f0b3c1ab21cb260b6a6732f8a6204b5aa0c90c5855bb36548: Status 404 returned error can't find the container with id f1c14ae3fe21779f0b3c1ab21cb260b6a6732f8a6204b5aa0c90c5855bb36548 Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.251203 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.468560 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.469223 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.469266 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"aded649576b848c24b0719426041f477f945986ac43e8c2293d6d810ad7dd3da"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.471050 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" event={"ID":"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8","Type":"ContainerStarted","Data":"c59c37c6cb4727a0a94ffe959b82047ee8f88c04a585c1c0b9edbc8a1efff05e"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.476290 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.476323 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ca1cfb105c5c4ea71ccf18e34164b2c90d2747660e20d45f974280080daf8cda"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.479972 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.480031 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"f1c14ae3fe21779f0b3c1ab21cb260b6a6732f8a6204b5aa0c90c5855bb36548"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.483699 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-k9pwr" event={"ID":"c07994f6-627f-4641-b164-ae81aac54782","Type":"ContainerStarted","Data":"d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.483760 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-k9pwr" event={"ID":"c07994f6-627f-4641-b164-ae81aac54782","Type":"ContainerStarted","Data":"35b4fd2baf714601a86109ecc1ac0179ffefdc14a2e3a605494d49c327cc3f88"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.486270 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.486295 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.486927 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.489684 4808 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb" exitCode=255 Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.489772 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.489816 4808 scope.go:117] "RemoveContainer" containerID="5e0d18f4230d3ec037e3339cd2f725d060f528a4bd013a28c0350d25e3715586" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.492897 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"675dbcd451a62be55a98b896f7263706148b09c294dbd88aa204d04296a08fdd"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.500910 4808 generic.go:334] "Generic (PLEG): container finished" podID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerID="4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09" exitCode=0 Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.500984 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerDied","Data":"4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.501043 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerStarted","Data":"e2e107b74a075760d5543d3e2da94b5776dcdda42690920b33b3ca2d8febdac2"} Nov 24 17:27:09 crc kubenswrapper[4808]: E1124 17:27:09.501912 4808 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.502085 4808 scope.go:117] "RemoveContainer" containerID="ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb" Nov 24 17:27:09 crc kubenswrapper[4808]: E1124 17:27:09.502231 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.504255 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q5m49" event={"ID":"6a192cd6-853d-4643-855d-ac61168b1cda","Type":"ContainerStarted","Data":"e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.504392 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q5m49" event={"ID":"6a192cd6-853d-4643-855d-ac61168b1cda","Type":"ContainerStarted","Data":"31d35c20e79f73aa4124bc9bc8da0f3ab2925eeb43022667178b66083469446f"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.507007 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7cbtg" event={"ID":"ded1d94f-dc75-45de-b921-a66747cd3f18","Type":"ContainerStarted","Data":"f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.507091 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7cbtg" event={"ID":"ded1d94f-dc75-45de-b921-a66747cd3f18","Type":"ContainerStarted","Data":"bd703f819dbe5dc21049b6f9c631a0f0391fab509cc2f03c51f4e2f6c01a3a62"} Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.507758 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.519821 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.533823 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.549085 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d18f4230d3ec037e3339cd2f725d060f528a4bd013a28c0350d25e3715586\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:02Z\\\",\\\"message\\\":\\\"W1124 17:26:51.512994 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 17:26:51.513507 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764005211 cert, and key in /tmp/serving-cert-2994416005/serving-signer.crt, /tmp/serving-cert-2994416005/serving-signer.key\\\\nI1124 17:26:51.887454 1 observer_polling.go:159] Starting file observer\\\\nW1124 17:26:51.890403 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 17:26:51.890599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:26:51.891348 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2994416005/tls.crt::/tmp/serving-cert-2994416005/tls.key\\\\\\\"\\\\nF1124 17:27:02.148785 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.565306 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.580573 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.602437 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.639103 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.671585 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.711086 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.755196 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.795062 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.832884 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.857521 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:27:09 crc kubenswrapper[4808]: E1124 17:27:09.857745 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:27:11.857699322 +0000 UTC m=+24.455367154 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.871830 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.910309 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.951206 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.958590 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.958755 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.958856 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:09 crc kubenswrapper[4808]: I1124 17:27:09.959010 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:09 crc kubenswrapper[4808]: E1124 17:27:09.958768 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:27:09 crc kubenswrapper[4808]: E1124 17:27:09.959324 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:11.959301422 +0000 UTC m=+24.556969234 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:27:09 crc kubenswrapper[4808]: E1124 17:27:09.958925 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:27:09 crc kubenswrapper[4808]: E1124 17:27:09.959520 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:27:09 crc kubenswrapper[4808]: E1124 17:27:09.959614 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:09 crc kubenswrapper[4808]: E1124 17:27:09.959771 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:11.959756705 +0000 UTC m=+24.557424507 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:09 crc kubenswrapper[4808]: E1124 17:27:09.959094 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:27:09 crc kubenswrapper[4808]: E1124 17:27:09.959937 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:11.95992847 +0000 UTC m=+24.557596272 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:27:09 crc kubenswrapper[4808]: E1124 17:27:09.959181 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:27:09 crc kubenswrapper[4808]: E1124 17:27:09.960091 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:27:09 crc kubenswrapper[4808]: E1124 17:27:09.960153 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:09 crc kubenswrapper[4808]: E1124 17:27:09.960236 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:11.960227498 +0000 UTC m=+24.557895300 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.000897 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.031946 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d18f4230d3ec037e3339cd2f725d060f528a4bd013a28c0350d25e3715586\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:02Z\\\",\\\"message\\\":\\\"W1124 17:26:51.512994 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 17:26:51.513507 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764005211 cert, and key in /tmp/serving-cert-2994416005/serving-signer.crt, /tmp/serving-cert-2994416005/serving-signer.key\\\\nI1124 17:26:51.887454 1 observer_polling.go:159] Starting file observer\\\\nW1124 17:26:51.890403 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 17:26:51.890599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:26:51.891348 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2994416005/tls.crt::/tmp/serving-cert-2994416005/tls.key\\\\\\\"\\\\nF1124 17:27:02.148785 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.069849 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.113970 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.156503 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.196596 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.237347 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.290595 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.310833 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.346476 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.346586 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:10 crc kubenswrapper[4808]: E1124 17:27:10.346878 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.346680 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:10 crc kubenswrapper[4808]: E1124 17:27:10.347314 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:10 crc kubenswrapper[4808]: E1124 17:27:10.347107 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.351815 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.353093 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.353866 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.354491 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.355205 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.355901 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.356558 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.357219 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.511824 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.515365 4808 scope.go:117] "RemoveContainer" containerID="ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb" Nov 24 17:27:10 crc kubenswrapper[4808]: E1124 17:27:10.515731 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.517050 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe"} Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.521125 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerStarted","Data":"e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481"} Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.521233 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerStarted","Data":"04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672"} Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.521329 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerStarted","Data":"793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d"} Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.521454 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerStarted","Data":"f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24"} Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.521535 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerStarted","Data":"920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a"} Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.521611 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerStarted","Data":"887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c"} Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.523441 4808 generic.go:334] "Generic (PLEG): container finished" podID="fe424e29-3d8f-49e3-ac0c-33dc63f93cf8" containerID="af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711" exitCode=0 Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.523578 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" event={"ID":"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8","Type":"ContainerDied","Data":"af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711"} Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.530170 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.543039 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.559876 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.580480 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.599640 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.617177 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.631719 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.648109 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.673067 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.738442 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.762849 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.845221 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.886487 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.916810 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.935937 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.952488 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:10 crc kubenswrapper[4808]: I1124 17:27:10.989702 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:10Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.032694 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.071654 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.111034 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.151190 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.196036 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.233644 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.276949 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.319658 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.355267 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.530248 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470"} Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.532616 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" event={"ID":"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8","Type":"ContainerStarted","Data":"7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc"} Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.547606 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.561167 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.570758 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.583773 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.600259 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.617264 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.628920 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.668892 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.712175 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.751754 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.795200 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.831479 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.882483 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:27:11 crc kubenswrapper[4808]: E1124 17:27:11.882743 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:27:15.882703334 +0000 UTC m=+28.480371136 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.890539 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.928293 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.953482 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:11Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.959802 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.971632 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.983676 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.983712 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.983749 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.983788 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:11 crc kubenswrapper[4808]: E1124 17:27:11.983912 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:27:11 crc kubenswrapper[4808]: E1124 17:27:11.983942 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:27:11 crc kubenswrapper[4808]: E1124 17:27:11.983958 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:11 crc kubenswrapper[4808]: E1124 17:27:11.983917 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:27:11 crc kubenswrapper[4808]: E1124 17:27:11.984010 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:15.983991594 +0000 UTC m=+28.581659396 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:11 crc kubenswrapper[4808]: E1124 17:27:11.984003 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:27:11 crc kubenswrapper[4808]: E1124 17:27:11.984194 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:27:11 crc kubenswrapper[4808]: E1124 17:27:11.984215 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:27:11 crc kubenswrapper[4808]: E1124 17:27:11.984228 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:11 crc kubenswrapper[4808]: E1124 17:27:11.984179 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:15.984151519 +0000 UTC m=+28.581819321 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:27:11 crc kubenswrapper[4808]: E1124 17:27:11.984283 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:15.984264352 +0000 UTC m=+28.581932154 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:27:11 crc kubenswrapper[4808]: E1124 17:27:11.984303 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:15.984292103 +0000 UTC m=+28.581959895 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:11 crc kubenswrapper[4808]: I1124 17:27:11.993722 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.014166 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.055582 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.089785 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.132047 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.171597 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.209469 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.249844 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.300922 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.346868 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:12 crc kubenswrapper[4808]: E1124 17:27:12.347038 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.347359 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.347517 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:12 crc kubenswrapper[4808]: E1124 17:27:12.347620 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:12 crc kubenswrapper[4808]: E1124 17:27:12.347680 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.357562 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.376773 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.411936 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.451955 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.510680 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.535641 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.538998 4808 generic.go:334] "Generic (PLEG): container finished" podID="fe424e29-3d8f-49e3-ac0c-33dc63f93cf8" containerID="7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc" exitCode=0 Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.539136 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" event={"ID":"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8","Type":"ContainerDied","Data":"7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc"} Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.575509 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.614731 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.655089 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.692406 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.730104 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.774242 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.813317 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.853708 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.891576 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.942106 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:12 crc kubenswrapper[4808]: I1124 17:27:12.977575 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.012006 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.053225 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.097572 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.139859 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.157458 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.158367 4808 scope.go:117] "RemoveContainer" containerID="ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb" Nov 24 17:27:13 crc kubenswrapper[4808]: E1124 17:27:13.158796 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.176788 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.215737 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.254452 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.293333 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.338302 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.389042 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.408705 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.450908 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.493503 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.532818 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.552670 4808 generic.go:334] "Generic (PLEG): container finished" podID="fe424e29-3d8f-49e3-ac0c-33dc63f93cf8" containerID="be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf" exitCode=0 Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.552786 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" event={"ID":"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8","Type":"ContainerDied","Data":"be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf"} Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.561706 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerStarted","Data":"cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee"} Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.569665 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.605698 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.611133 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.613135 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.631345 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.672734 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.709422 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.750754 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.796482 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.831631 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.874649 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.910496 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.954623 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:13 crc kubenswrapper[4808]: I1124 17:27:13.993843 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.033905 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.076773 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.113761 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.151216 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.199395 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.228425 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.231291 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.231339 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.231351 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.231498 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.237054 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.283938 4808 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.284518 4808 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.285607 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.285658 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.285673 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.285690 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.285699 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:14Z","lastTransitionTime":"2025-11-24T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:14 crc kubenswrapper[4808]: E1124 17:27:14.299168 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.302572 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.302599 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.302625 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.302641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.302649 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:14Z","lastTransitionTime":"2025-11-24T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.319617 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: E1124 17:27:14.320276 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.324112 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.324154 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.324166 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.324412 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.324432 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:14Z","lastTransitionTime":"2025-11-24T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:14 crc kubenswrapper[4808]: E1124 17:27:14.336597 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.341852 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.342209 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.342322 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.342394 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.342499 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:14Z","lastTransitionTime":"2025-11-24T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.346926 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.347009 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.347067 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:14 crc kubenswrapper[4808]: E1124 17:27:14.347286 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:14 crc kubenswrapper[4808]: E1124 17:27:14.347388 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:14 crc kubenswrapper[4808]: E1124 17:27:14.347271 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.351626 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: E1124 17:27:14.357929 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.360855 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.360899 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.360908 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.360922 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.360932 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:14Z","lastTransitionTime":"2025-11-24T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:14 crc kubenswrapper[4808]: E1124 17:27:14.372228 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: E1124 17:27:14.372558 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.374293 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.374320 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.374329 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.374343 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.374353 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:14Z","lastTransitionTime":"2025-11-24T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.389427 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.430155 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.472906 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.476870 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.476936 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.476947 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.476967 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.476979 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:14Z","lastTransitionTime":"2025-11-24T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.511642 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.550293 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.567450 4808 generic.go:334] "Generic (PLEG): container finished" podID="fe424e29-3d8f-49e3-ac0c-33dc63f93cf8" containerID="f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8" exitCode=0 Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.567564 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" event={"ID":"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8","Type":"ContainerDied","Data":"f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8"} Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.578661 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.578704 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.578714 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.578732 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.578748 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:14Z","lastTransitionTime":"2025-11-24T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.592415 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.635867 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.672752 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.682486 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.682533 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.682568 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.682585 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.682616 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:14Z","lastTransitionTime":"2025-11-24T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.711722 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.755210 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.785899 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.785940 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.785951 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.785971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.785983 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:14Z","lastTransitionTime":"2025-11-24T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.791584 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.831757 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.888337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.888482 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.888494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.888519 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.888534 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:14Z","lastTransitionTime":"2025-11-24T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.889855 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.911232 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.953620 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.990999 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.991748 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.991809 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.991829 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.991855 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:14 crc kubenswrapper[4808]: I1124 17:27:14.991873 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:14Z","lastTransitionTime":"2025-11-24T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.031528 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.076142 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.095173 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.095201 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.095211 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.095227 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.095237 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:15Z","lastTransitionTime":"2025-11-24T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.110451 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.160293 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.198412 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.198479 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.198496 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.198523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.198544 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:15Z","lastTransitionTime":"2025-11-24T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.203052 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.231710 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.270821 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.301482 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.301539 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.301555 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.301582 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.301599 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:15Z","lastTransitionTime":"2025-11-24T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.311794 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.356510 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.404187 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.404246 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.404259 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.404281 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.404292 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:15Z","lastTransitionTime":"2025-11-24T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.507449 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.507531 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.507597 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.507617 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.507627 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:15Z","lastTransitionTime":"2025-11-24T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.575654 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerStarted","Data":"5f5ffc56c431d311e8e0af9df3a1b137b35267c135cd69982514429cf328cdbb"} Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.576077 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.579411 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" event={"ID":"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8","Type":"ContainerStarted","Data":"83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93"} Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.590780 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.602986 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.611512 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.611566 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.611580 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.611598 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.611611 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:15Z","lastTransitionTime":"2025-11-24T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.621219 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5ffc56c431d311e8e0af9df3a1b137b35267c135cd69982514429cf328cdbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.633463 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.641657 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.655375 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.670653 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.681364 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.693519 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.711626 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.714759 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.714811 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.714823 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.714849 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.714863 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:15Z","lastTransitionTime":"2025-11-24T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.752744 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.818284 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.818345 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.818374 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.818404 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.818423 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:15Z","lastTransitionTime":"2025-11-24T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.822170 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.847260 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.878211 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.910800 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.920969 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.921040 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.921054 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.921075 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.921090 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:15Z","lastTransitionTime":"2025-11-24T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.934838 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:27:15 crc kubenswrapper[4808]: E1124 17:27:15.935146 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:27:23.935114017 +0000 UTC m=+36.532781819 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.954812 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:15 crc kubenswrapper[4808]: I1124 17:27:15.989678 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.025138 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.025204 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.025222 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.025242 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.025257 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:16Z","lastTransitionTime":"2025-11-24T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.029563 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.036006 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.036094 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.036126 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.036150 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:16 crc kubenswrapper[4808]: E1124 17:27:16.036240 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:27:16 crc kubenswrapper[4808]: E1124 17:27:16.036300 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:24.036279964 +0000 UTC m=+36.633947766 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:27:16 crc kubenswrapper[4808]: E1124 17:27:16.036309 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:27:16 crc kubenswrapper[4808]: E1124 17:27:16.036241 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:27:16 crc kubenswrapper[4808]: E1124 17:27:16.036471 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:27:16 crc kubenswrapper[4808]: E1124 17:27:16.036520 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:24.036416198 +0000 UTC m=+36.634084000 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:27:16 crc kubenswrapper[4808]: E1124 17:27:16.036533 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:27:16 crc kubenswrapper[4808]: E1124 17:27:16.036478 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:27:16 crc kubenswrapper[4808]: E1124 17:27:16.036593 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:16 crc kubenswrapper[4808]: E1124 17:27:16.036648 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:16 crc kubenswrapper[4808]: E1124 17:27:16.036657 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:24.036650215 +0000 UTC m=+36.634318017 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:16 crc kubenswrapper[4808]: E1124 17:27:16.036748 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:24.036731697 +0000 UTC m=+36.634399499 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.071810 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.111536 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.128042 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.128083 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.128094 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.128108 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.128117 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:16Z","lastTransitionTime":"2025-11-24T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.152954 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.198785 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.230987 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.231574 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.231613 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.231623 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.231641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.231657 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:16Z","lastTransitionTime":"2025-11-24T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.271950 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.312939 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.333972 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.334051 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.334066 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.334092 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.334106 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:16Z","lastTransitionTime":"2025-11-24T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.346774 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.346774 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.346800 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:16 crc kubenswrapper[4808]: E1124 17:27:16.346944 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:16 crc kubenswrapper[4808]: E1124 17:27:16.347059 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:16 crc kubenswrapper[4808]: E1124 17:27:16.347133 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.355794 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.393133 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.432326 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.437653 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.437687 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.437700 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.437724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.437742 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:16Z","lastTransitionTime":"2025-11-24T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.480711 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5ffc56c431d311e8e0af9df3a1b137b35267c135cd69982514429cf328cdbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.514545 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.540177 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.540239 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.540254 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.540271 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.540281 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:16Z","lastTransitionTime":"2025-11-24T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.554187 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.585277 4808 generic.go:334] "Generic (PLEG): container finished" podID="fe424e29-3d8f-49e3-ac0c-33dc63f93cf8" containerID="83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93" exitCode=0 Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.585392 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" event={"ID":"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8","Type":"ContainerDied","Data":"83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93"} Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.585501 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.585430 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.603843 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.615791 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.630787 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.643547 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.643600 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.643609 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.643624 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.643633 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:16Z","lastTransitionTime":"2025-11-24T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.679333 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5ffc56c431d311e8e0af9df3a1b137b35267c135cd69982514429cf328cdbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.716988 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.746962 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.747007 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.747033 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.747049 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.747060 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:16Z","lastTransitionTime":"2025-11-24T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.751082 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.792651 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.830915 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.851577 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.851610 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.851620 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.851637 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.851647 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:16Z","lastTransitionTime":"2025-11-24T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.871804 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.913635 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.951658 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.955234 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.955265 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.955275 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.955292 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.955303 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:16Z","lastTransitionTime":"2025-11-24T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:16 crc kubenswrapper[4808]: I1124 17:27:16.995376 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.033352 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.058622 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.058675 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.058689 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.058710 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.058724 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:17Z","lastTransitionTime":"2025-11-24T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.070592 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.111216 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.149677 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.162354 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.162422 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.162435 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.162450 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.162460 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:17Z","lastTransitionTime":"2025-11-24T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.191371 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.239510 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.264805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.264844 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.264853 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.264867 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.264876 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:17Z","lastTransitionTime":"2025-11-24T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.272986 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.310459 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.352768 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.367534 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.367589 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.367599 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.367621 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.367635 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:17Z","lastTransitionTime":"2025-11-24T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.393110 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.431409 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.470620 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.470689 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.470702 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.470726 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.470742 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:17Z","lastTransitionTime":"2025-11-24T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.473984 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.510304 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.550261 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.573420 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.573478 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.573494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.573517 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.573532 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:17Z","lastTransitionTime":"2025-11-24T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.593632 4808 generic.go:334] "Generic (PLEG): container finished" podID="fe424e29-3d8f-49e3-ac0c-33dc63f93cf8" containerID="8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8" exitCode=0 Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.593695 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" event={"ID":"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8","Type":"ContainerDied","Data":"8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8"} Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.593803 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.593810 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.632911 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.673885 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.675818 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.675858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.675871 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.675887 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.675897 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:17Z","lastTransitionTime":"2025-11-24T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.717748 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5ffc56c431d311e8e0af9df3a1b137b35267c135cd69982514429cf328cdbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.752452 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.778768 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.778824 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.778836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.778857 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.778873 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:17Z","lastTransitionTime":"2025-11-24T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.798402 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.834003 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.873872 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.882067 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.882117 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.882128 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.882152 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.882164 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:17Z","lastTransitionTime":"2025-11-24T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.917352 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5ffc56c431d311e8e0af9df3a1b137b35267c135cd69982514429cf328cdbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.951840 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.984082 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.984125 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.984139 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.984155 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.984165 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:17Z","lastTransitionTime":"2025-11-24T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:17 crc kubenswrapper[4808]: I1124 17:27:17.990969 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.039329 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.097559 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.097612 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.097625 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.097640 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.097650 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:18Z","lastTransitionTime":"2025-11-24T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.098417 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.110515 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.153158 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.192410 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.200181 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.200247 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.200263 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.200285 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.200300 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:18Z","lastTransitionTime":"2025-11-24T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.231495 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.272060 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.302664 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.302694 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.302704 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.302717 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.302727 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:18Z","lastTransitionTime":"2025-11-24T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.310746 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.348305 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.348339 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.348383 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:18 crc kubenswrapper[4808]: E1124 17:27:18.348442 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:18 crc kubenswrapper[4808]: E1124 17:27:18.348684 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:18 crc kubenswrapper[4808]: E1124 17:27:18.348582 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.359561 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.391626 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.409576 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.409626 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.409638 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.409654 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.409665 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:18Z","lastTransitionTime":"2025-11-24T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.442108 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.474423 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.509770 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.511740 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.511794 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.511807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.511824 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.511835 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:18Z","lastTransitionTime":"2025-11-24T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.552851 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.595344 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.599306 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovnkube-controller/0.log" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.603118 4808 generic.go:334] "Generic (PLEG): container finished" podID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerID="5f5ffc56c431d311e8e0af9df3a1b137b35267c135cd69982514429cf328cdbb" exitCode=1 Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.603179 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerDied","Data":"5f5ffc56c431d311e8e0af9df3a1b137b35267c135cd69982514429cf328cdbb"} Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.604395 4808 scope.go:117] "RemoveContainer" containerID="5f5ffc56c431d311e8e0af9df3a1b137b35267c135cd69982514429cf328cdbb" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.607491 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" event={"ID":"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8","Type":"ContainerStarted","Data":"6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69"} Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.614055 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.614102 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.614115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.614133 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.614147 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:18Z","lastTransitionTime":"2025-11-24T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.638803 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.674495 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.710733 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.720826 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.720876 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.720890 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.720910 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.720923 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:18Z","lastTransitionTime":"2025-11-24T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.753436 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.796598 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.827068 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.827151 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.827169 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.827205 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.827226 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:18Z","lastTransitionTime":"2025-11-24T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.834036 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.872631 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.930805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.930852 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.930866 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.930885 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.930898 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:18Z","lastTransitionTime":"2025-11-24T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.953769 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:18 crc kubenswrapper[4808]: I1124 17:27:18.993030 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5ffc56c431d311e8e0af9df3a1b137b35267c135cd69982514429cf328cdbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.034364 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.034415 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.034427 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.034445 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.034457 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:19Z","lastTransitionTime":"2025-11-24T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.034796 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5ffc56c431d311e8e0af9df3a1b137b35267c135cd69982514429cf328cdbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5ffc56c431d311e8e0af9df3a1b137b35267c135cd69982514429cf328cdbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"message\\\":\\\"nformers/externalversions/factory.go:140\\\\nI1124 17:27:18.446629 6104 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 17:27:18.447091 6104 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 17:27:18.447118 6104 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 17:27:18.447124 6104 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 17:27:18.447143 6104 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:27:18.447365 6104 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:27:18.447397 6104 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 17:27:18.447399 6104 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:27:18.447407 6104 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 17:27:18.447419 6104 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 17:27:18.447453 6104 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 17:27:18.447456 6104 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 17:27:18.447470 6104 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 17:27:18.447454 6104 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.051877 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.074341 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.111068 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.177538 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.177585 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.177594 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.177616 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.177628 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:19Z","lastTransitionTime":"2025-11-24T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.190176 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.205028 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.234196 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.272470 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.280120 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.280176 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.280190 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.280210 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.280225 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:19Z","lastTransitionTime":"2025-11-24T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.321231 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.352838 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.382823 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.382872 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.382881 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.382898 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.382908 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:19Z","lastTransitionTime":"2025-11-24T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.391224 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.438101 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.475731 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.486351 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.486393 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.486406 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.486424 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.486433 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:19Z","lastTransitionTime":"2025-11-24T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.511573 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.551839 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.589006 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.589061 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.589070 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.589084 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.589093 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:19Z","lastTransitionTime":"2025-11-24T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.613523 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovnkube-controller/1.log" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.614395 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovnkube-controller/0.log" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.618651 4808 generic.go:334] "Generic (PLEG): container finished" podID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerID="bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9" exitCode=1 Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.618799 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerDied","Data":"bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9"} Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.618901 4808 scope.go:117] "RemoveContainer" containerID="5f5ffc56c431d311e8e0af9df3a1b137b35267c135cd69982514429cf328cdbb" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.619775 4808 scope.go:117] "RemoveContainer" containerID="bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9" Nov 24 17:27:19 crc kubenswrapper[4808]: E1124 17:27:19.620070 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.637486 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.652918 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.672118 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.691689 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.691729 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.691749 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.691763 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.691773 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:19Z","lastTransitionTime":"2025-11-24T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.713300 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.751048 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.795331 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.795388 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.795399 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.795419 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.795434 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:19Z","lastTransitionTime":"2025-11-24T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.810205 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.837813 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.873988 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.898628 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.898685 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.898695 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.898713 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.898723 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:19Z","lastTransitionTime":"2025-11-24T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.918846 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.955506 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:19 crc kubenswrapper[4808]: I1124 17:27:19.990201 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.001041 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.001087 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.001099 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.001112 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.001123 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:20Z","lastTransitionTime":"2025-11-24T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.031349 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.076410 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5ffc56c431d311e8e0af9df3a1b137b35267c135cd69982514429cf328cdbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"message\\\":\\\"nformers/externalversions/factory.go:140\\\\nI1124 17:27:18.446629 6104 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 17:27:18.447091 6104 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 17:27:18.447118 6104 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 17:27:18.447124 6104 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 17:27:18.447143 6104 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:27:18.447365 6104 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:27:18.447397 6104 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 17:27:18.447399 6104 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:27:18.447407 6104 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 17:27:18.447419 6104 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 17:27:18.447453 6104 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 17:27:18.447456 6104 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 17:27:18.447470 6104 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 17:27:18.447454 6104 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:19Z\\\",\\\"message\\\":\\\" after 0 failed attempt(s)\\\\nI1124 17:27:19.478433 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1124 17:27:19.478436 6277 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1124 17:27:19.478438 6277 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:19.478416 6277 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1124 17:27:19.478439 6277 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1124 17:27:19.478380 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-q5m49 after 0 failed attempt(s)\\\\nI1124 17:27:19.478484 6277 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-q5m49\\\\nF1124 17:27:19.478364 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.103940 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.104058 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.104079 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.104103 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.104118 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:20Z","lastTransitionTime":"2025-11-24T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.127562 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.154664 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.207088 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.207131 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.207140 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.207158 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.207170 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:20Z","lastTransitionTime":"2025-11-24T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.310264 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.310309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.310318 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.310338 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.310350 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:20Z","lastTransitionTime":"2025-11-24T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.347216 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.347246 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.347216 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:20 crc kubenswrapper[4808]: E1124 17:27:20.347401 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:20 crc kubenswrapper[4808]: E1124 17:27:20.347522 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:20 crc kubenswrapper[4808]: E1124 17:27:20.347734 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.412829 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.412884 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.412895 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.412913 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.412926 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:20Z","lastTransitionTime":"2025-11-24T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.515969 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.516048 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.516065 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.516084 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.516098 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:20Z","lastTransitionTime":"2025-11-24T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.619510 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.619572 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.619585 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.619607 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.619626 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:20Z","lastTransitionTime":"2025-11-24T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.623803 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovnkube-controller/1.log" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.628078 4808 scope.go:117] "RemoveContainer" containerID="bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9" Nov 24 17:27:20 crc kubenswrapper[4808]: E1124 17:27:20.628426 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.640734 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.652120 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.667322 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.681180 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.693580 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.705714 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.716418 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.721884 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.721930 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.721944 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.721960 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.721971 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:20Z","lastTransitionTime":"2025-11-24T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.734971 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:19Z\\\",\\\"message\\\":\\\" after 0 failed attempt(s)\\\\nI1124 17:27:19.478433 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1124 17:27:19.478436 6277 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1124 17:27:19.478438 6277 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:19.478416 6277 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1124 17:27:19.478439 6277 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1124 17:27:19.478380 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-q5m49 after 0 failed attempt(s)\\\\nI1124 17:27:19.478484 6277 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-q5m49\\\\nF1124 17:27:19.478364 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.748226 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.767178 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.778809 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.788068 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.800683 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.813165 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.824451 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.824502 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.824513 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.824528 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.824536 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:20Z","lastTransitionTime":"2025-11-24T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.827674 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.926974 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.927042 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.927056 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.927073 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:20 crc kubenswrapper[4808]: I1124 17:27:20.927084 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:20Z","lastTransitionTime":"2025-11-24T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.030030 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.030081 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.030094 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.030113 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.030126 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:21Z","lastTransitionTime":"2025-11-24T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.133606 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.133652 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.133664 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.133683 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.133695 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:21Z","lastTransitionTime":"2025-11-24T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.236263 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.236299 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.236318 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.236335 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.236345 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:21Z","lastTransitionTime":"2025-11-24T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.340176 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.340240 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.340258 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.340283 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.340299 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:21Z","lastTransitionTime":"2025-11-24T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.442878 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.442944 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.442962 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.442987 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.443006 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:21Z","lastTransitionTime":"2025-11-24T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.502535 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b"] Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.503100 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.507185 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.507333 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.530598 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.546687 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.546768 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.546793 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.546826 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.546849 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:21Z","lastTransitionTime":"2025-11-24T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.555055 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.571313 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.586170 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.602439 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.603408 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k22s4\" (UniqueName: \"kubernetes.io/projected/8ce6613f-d181-4eb4-9cfc-4762964eb8d0-kube-api-access-k22s4\") pod \"ovnkube-control-plane-749d76644c-p5b8b\" (UID: \"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.603477 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8ce6613f-d181-4eb4-9cfc-4762964eb8d0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p5b8b\" (UID: \"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.603522 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8ce6613f-d181-4eb4-9cfc-4762964eb8d0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p5b8b\" (UID: \"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.603643 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8ce6613f-d181-4eb4-9cfc-4762964eb8d0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p5b8b\" (UID: \"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.624951 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.642147 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.650689 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.650768 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.650789 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.650821 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.650846 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:21Z","lastTransitionTime":"2025-11-24T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.663714 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.690064 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:19Z\\\",\\\"message\\\":\\\" after 0 failed attempt(s)\\\\nI1124 17:27:19.478433 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1124 17:27:19.478436 6277 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1124 17:27:19.478438 6277 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:19.478416 6277 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1124 17:27:19.478439 6277 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1124 17:27:19.478380 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-q5m49 after 0 failed attempt(s)\\\\nI1124 17:27:19.478484 6277 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-q5m49\\\\nF1124 17:27:19.478364 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.705117 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k22s4\" (UniqueName: \"kubernetes.io/projected/8ce6613f-d181-4eb4-9cfc-4762964eb8d0-kube-api-access-k22s4\") pod \"ovnkube-control-plane-749d76644c-p5b8b\" (UID: \"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.705162 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8ce6613f-d181-4eb4-9cfc-4762964eb8d0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p5b8b\" (UID: \"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.705184 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8ce6613f-d181-4eb4-9cfc-4762964eb8d0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p5b8b\" (UID: \"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.705238 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8ce6613f-d181-4eb4-9cfc-4762964eb8d0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p5b8b\" (UID: \"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.706858 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8ce6613f-d181-4eb4-9cfc-4762964eb8d0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p5b8b\" (UID: \"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.707243 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8ce6613f-d181-4eb4-9cfc-4762964eb8d0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p5b8b\" (UID: \"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.709977 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.714222 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8ce6613f-d181-4eb4-9cfc-4762964eb8d0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p5b8b\" (UID: \"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.722106 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k22s4\" (UniqueName: \"kubernetes.io/projected/8ce6613f-d181-4eb4-9cfc-4762964eb8d0-kube-api-access-k22s4\") pod \"ovnkube-control-plane-749d76644c-p5b8b\" (UID: \"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.724104 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.746191 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.754839 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.754935 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.754963 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.754986 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.754999 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:21Z","lastTransitionTime":"2025-11-24T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.768001 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.784418 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.804450 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.818955 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.819430 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.866836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.866976 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.867453 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.867474 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.867539 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:21Z","lastTransitionTime":"2025-11-24T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.970476 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.970523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.970533 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.970554 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:21 crc kubenswrapper[4808]: I1124 17:27:21.970566 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:21Z","lastTransitionTime":"2025-11-24T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.075337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.075369 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.075378 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.075391 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.075400 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:22Z","lastTransitionTime":"2025-11-24T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.178595 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.178626 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.178635 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.178649 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.178658 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:22Z","lastTransitionTime":"2025-11-24T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.255566 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-8v2q4"] Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.256147 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:22 crc kubenswrapper[4808]: E1124 17:27:22.256216 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.272808 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.280667 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.280706 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.280715 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.280731 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.280742 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:22Z","lastTransitionTime":"2025-11-24T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.291970 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.311396 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.325232 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.335390 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.346916 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.346866 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.346998 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.347049 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:22 crc kubenswrapper[4808]: E1124 17:27:22.347141 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:22 crc kubenswrapper[4808]: E1124 17:27:22.347225 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:22 crc kubenswrapper[4808]: E1124 17:27:22.347360 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.360041 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.372884 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.383580 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.383622 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.383631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.383645 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.383656 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:22Z","lastTransitionTime":"2025-11-24T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.390629 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:19Z\\\",\\\"message\\\":\\\" after 0 failed attempt(s)\\\\nI1124 17:27:19.478433 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1124 17:27:19.478436 6277 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1124 17:27:19.478438 6277 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:19.478416 6277 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1124 17:27:19.478439 6277 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1124 17:27:19.478380 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-q5m49 after 0 failed attempt(s)\\\\nI1124 17:27:19.478484 6277 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-q5m49\\\\nF1124 17:27:19.478364 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.403164 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.412353 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs\") pod \"network-metrics-daemon-8v2q4\" (UID: \"c26458eb-7f6c-42cb-b6c5-90de6756a994\") " pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.412433 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkr64\" (UniqueName: \"kubernetes.io/projected/c26458eb-7f6c-42cb-b6c5-90de6756a994-kube-api-access-xkr64\") pod \"network-metrics-daemon-8v2q4\" (UID: \"c26458eb-7f6c-42cb-b6c5-90de6756a994\") " pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.417249 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.429672 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.442812 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.464390 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.477371 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.485425 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.485462 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.485472 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.485486 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.485498 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:22Z","lastTransitionTime":"2025-11-24T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.489119 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.501398 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.514069 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkr64\" (UniqueName: \"kubernetes.io/projected/c26458eb-7f6c-42cb-b6c5-90de6756a994-kube-api-access-xkr64\") pod \"network-metrics-daemon-8v2q4\" (UID: \"c26458eb-7f6c-42cb-b6c5-90de6756a994\") " pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.514218 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs\") pod \"network-metrics-daemon-8v2q4\" (UID: \"c26458eb-7f6c-42cb-b6c5-90de6756a994\") " pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:22 crc kubenswrapper[4808]: E1124 17:27:22.514358 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:27:22 crc kubenswrapper[4808]: E1124 17:27:22.514439 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs podName:c26458eb-7f6c-42cb-b6c5-90de6756a994 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:23.014415425 +0000 UTC m=+35.612083247 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs") pod "network-metrics-daemon-8v2q4" (UID: "c26458eb-7f6c-42cb-b6c5-90de6756a994") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.529379 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkr64\" (UniqueName: \"kubernetes.io/projected/c26458eb-7f6c-42cb-b6c5-90de6756a994-kube-api-access-xkr64\") pod \"network-metrics-daemon-8v2q4\" (UID: \"c26458eb-7f6c-42cb-b6c5-90de6756a994\") " pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.587460 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.587517 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.587534 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.587557 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.587577 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:22Z","lastTransitionTime":"2025-11-24T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.636536 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" event={"ID":"8ce6613f-d181-4eb4-9cfc-4762964eb8d0","Type":"ContainerStarted","Data":"da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe"} Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.636626 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" event={"ID":"8ce6613f-d181-4eb4-9cfc-4762964eb8d0","Type":"ContainerStarted","Data":"cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24"} Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.636658 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" event={"ID":"8ce6613f-d181-4eb4-9cfc-4762964eb8d0","Type":"ContainerStarted","Data":"f813e393093ee9d7a153e082a9fb6db337a6b23e65d5ccbc111e71e2980e6197"} Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.651167 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.678431 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.690063 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.690112 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.690128 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.690152 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.690170 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:22Z","lastTransitionTime":"2025-11-24T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.691353 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.702663 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.717724 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.730874 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.744990 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.756983 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.769539 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.779626 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.790620 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.792160 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.792185 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.792196 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.792212 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.792223 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:22Z","lastTransitionTime":"2025-11-24T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.801275 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.814229 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.826316 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.835894 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.854187 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:19Z\\\",\\\"message\\\":\\\" after 0 failed attempt(s)\\\\nI1124 17:27:19.478433 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1124 17:27:19.478436 6277 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1124 17:27:19.478438 6277 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:19.478416 6277 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1124 17:27:19.478439 6277 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1124 17:27:19.478380 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-q5m49 after 0 failed attempt(s)\\\\nI1124 17:27:19.478484 6277 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-q5m49\\\\nF1124 17:27:19.478364 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.868214 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.894045 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.894093 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.894104 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.894121 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.894133 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:22Z","lastTransitionTime":"2025-11-24T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.997683 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.997743 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.997761 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.997788 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:22 crc kubenswrapper[4808]: I1124 17:27:22.997808 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:22Z","lastTransitionTime":"2025-11-24T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.019352 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs\") pod \"network-metrics-daemon-8v2q4\" (UID: \"c26458eb-7f6c-42cb-b6c5-90de6756a994\") " pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:23 crc kubenswrapper[4808]: E1124 17:27:23.019567 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:27:23 crc kubenswrapper[4808]: E1124 17:27:23.019697 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs podName:c26458eb-7f6c-42cb-b6c5-90de6756a994 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:24.019675804 +0000 UTC m=+36.617343606 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs") pod "network-metrics-daemon-8v2q4" (UID: "c26458eb-7f6c-42cb-b6c5-90de6756a994") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.100536 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.100569 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.100580 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.100595 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.100604 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:23Z","lastTransitionTime":"2025-11-24T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.203480 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.203587 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.203605 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.203671 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.203691 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:23Z","lastTransitionTime":"2025-11-24T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.307175 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.307221 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.307232 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.307249 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.307260 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:23Z","lastTransitionTime":"2025-11-24T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.409412 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.409453 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.409462 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.409475 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.409484 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:23Z","lastTransitionTime":"2025-11-24T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.512877 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.512923 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.512933 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.512949 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.512964 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:23Z","lastTransitionTime":"2025-11-24T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.615131 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.615196 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.615207 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.615223 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.615233 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:23Z","lastTransitionTime":"2025-11-24T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.718523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.718589 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.718603 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.718628 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.718643 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:23Z","lastTransitionTime":"2025-11-24T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.822188 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.822656 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.822699 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.822726 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.822740 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:23Z","lastTransitionTime":"2025-11-24T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.926805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.926877 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.926889 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.926938 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:23 crc kubenswrapper[4808]: I1124 17:27:23.926957 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:23Z","lastTransitionTime":"2025-11-24T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.029804 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.030002 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:27:40.029958447 +0000 UTC m=+52.627626249 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.030201 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs\") pod \"network-metrics-daemon-8v2q4\" (UID: \"c26458eb-7f6c-42cb-b6c5-90de6756a994\") " pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.030524 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.030660 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs podName:c26458eb-7f6c-42cb-b6c5-90de6756a994 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:26.030630556 +0000 UTC m=+38.628298368 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs") pod "network-metrics-daemon-8v2q4" (UID: "c26458eb-7f6c-42cb-b6c5-90de6756a994") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.030958 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.032863 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.032913 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.033033 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.033053 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:24Z","lastTransitionTime":"2025-11-24T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.131860 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.131940 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.131977 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.132099 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.132154 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.132282 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.132293 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.132320 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:40.132291898 +0000 UTC m=+52.729959710 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.132329 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.132342 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:40.132331379 +0000 UTC m=+52.729999181 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.132350 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.132372 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.132405 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.132418 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.132446 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:40.132419451 +0000 UTC m=+52.730087283 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.132479 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:40.132458833 +0000 UTC m=+52.730126635 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.136833 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.136919 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.136952 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.137083 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.137104 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:24Z","lastTransitionTime":"2025-11-24T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.240746 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.240834 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.240857 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.240889 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.240912 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:24Z","lastTransitionTime":"2025-11-24T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.343593 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.343671 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.343688 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.343715 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.343734 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:24Z","lastTransitionTime":"2025-11-24T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.347084 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.347146 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.347154 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.347383 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.347407 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.347538 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.347529 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.347608 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.348507 4808 scope.go:117] "RemoveContainer" containerID="ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.445911 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.445961 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.445974 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.445991 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.446002 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:24Z","lastTransitionTime":"2025-11-24T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.460806 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.467794 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.467838 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.467847 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.467862 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.467875 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:24Z","lastTransitionTime":"2025-11-24T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.483301 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.488724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.488762 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.488773 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.488795 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.488811 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:24Z","lastTransitionTime":"2025-11-24T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.502491 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.507912 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.507958 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.507967 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.507985 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.507997 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:24Z","lastTransitionTime":"2025-11-24T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.522529 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.526976 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.527035 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.527058 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.527087 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.527100 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:24Z","lastTransitionTime":"2025-11-24T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.541700 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: E1124 17:27:24.541836 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.545342 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.545391 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.545404 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.545422 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.545439 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:24Z","lastTransitionTime":"2025-11-24T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.646606 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.647343 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.647554 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.647631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.647654 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.647664 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:24Z","lastTransitionTime":"2025-11-24T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.649129 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9"} Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.649597 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.666160 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.679776 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.691657 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.711312 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:19Z\\\",\\\"message\\\":\\\" after 0 failed attempt(s)\\\\nI1124 17:27:19.478433 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1124 17:27:19.478436 6277 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1124 17:27:19.478438 6277 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:19.478416 6277 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1124 17:27:19.478439 6277 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1124 17:27:19.478380 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-q5m49 after 0 failed attempt(s)\\\\nI1124 17:27:19.478484 6277 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-q5m49\\\\nF1124 17:27:19.478364 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.725483 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.740885 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.749587 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.749616 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.749625 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.749641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.749651 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:24Z","lastTransitionTime":"2025-11-24T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.751961 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.765830 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.784556 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.797934 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.810219 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.823353 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.838139 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.851984 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.852037 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.852047 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.852062 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.852071 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:24Z","lastTransitionTime":"2025-11-24T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.853716 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.864919 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.876864 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.885861 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.955492 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.955554 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.955567 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.955589 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:24 crc kubenswrapper[4808]: I1124 17:27:24.955604 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:24Z","lastTransitionTime":"2025-11-24T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.059181 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.059281 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.059316 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.059351 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.059384 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:25Z","lastTransitionTime":"2025-11-24T17:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.162254 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.162311 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.162331 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.162358 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.162372 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:25Z","lastTransitionTime":"2025-11-24T17:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.265294 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.265344 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.265354 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.265370 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.265380 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:25Z","lastTransitionTime":"2025-11-24T17:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.368671 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.368728 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.368741 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.368768 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.368780 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:25Z","lastTransitionTime":"2025-11-24T17:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.471739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.471797 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.471811 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.471831 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.471846 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:25Z","lastTransitionTime":"2025-11-24T17:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.574886 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.574980 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.575004 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.575073 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.575099 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:25Z","lastTransitionTime":"2025-11-24T17:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.677889 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.677925 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.677933 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.677947 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.677956 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:25Z","lastTransitionTime":"2025-11-24T17:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.780833 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.780862 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.780871 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.780884 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.780893 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:25Z","lastTransitionTime":"2025-11-24T17:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.884149 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.884228 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.884244 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.884266 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.884283 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:25Z","lastTransitionTime":"2025-11-24T17:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.987413 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.987450 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.987459 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.987472 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:25 crc kubenswrapper[4808]: I1124 17:27:25.987480 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:25Z","lastTransitionTime":"2025-11-24T17:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.056255 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs\") pod \"network-metrics-daemon-8v2q4\" (UID: \"c26458eb-7f6c-42cb-b6c5-90de6756a994\") " pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:26 crc kubenswrapper[4808]: E1124 17:27:26.056477 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:27:26 crc kubenswrapper[4808]: E1124 17:27:26.056566 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs podName:c26458eb-7f6c-42cb-b6c5-90de6756a994 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:30.056547134 +0000 UTC m=+42.654214936 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs") pod "network-metrics-daemon-8v2q4" (UID: "c26458eb-7f6c-42cb-b6c5-90de6756a994") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.090066 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.090126 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.090135 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.090153 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.090166 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:26Z","lastTransitionTime":"2025-11-24T17:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.192198 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.192259 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.192273 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.192293 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.192306 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:26Z","lastTransitionTime":"2025-11-24T17:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.294741 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.294813 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.294822 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.294836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.294846 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:26Z","lastTransitionTime":"2025-11-24T17:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.346392 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.346545 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:26 crc kubenswrapper[4808]: E1124 17:27:26.346599 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.346652 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.346666 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:26 crc kubenswrapper[4808]: E1124 17:27:26.346756 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:26 crc kubenswrapper[4808]: E1124 17:27:26.346968 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:26 crc kubenswrapper[4808]: E1124 17:27:26.347101 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.397273 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.397321 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.397334 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.397354 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.397367 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:26Z","lastTransitionTime":"2025-11-24T17:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.499982 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.500027 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.500037 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.500051 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.500065 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:26Z","lastTransitionTime":"2025-11-24T17:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.602677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.602758 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.602777 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.602808 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.602827 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:26Z","lastTransitionTime":"2025-11-24T17:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.705647 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.705729 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.705753 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.705790 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.705817 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:26Z","lastTransitionTime":"2025-11-24T17:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.809376 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.809443 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.809461 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.809494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.809513 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:26Z","lastTransitionTime":"2025-11-24T17:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.912188 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.912258 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.912269 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.912292 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:26 crc kubenswrapper[4808]: I1124 17:27:26.912305 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:26Z","lastTransitionTime":"2025-11-24T17:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.014919 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.014993 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.015009 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.015053 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.015069 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:27Z","lastTransitionTime":"2025-11-24T17:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.118116 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.118155 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.118173 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.118193 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.118204 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:27Z","lastTransitionTime":"2025-11-24T17:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.221188 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.221263 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.221276 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.221303 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.221321 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:27Z","lastTransitionTime":"2025-11-24T17:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.324255 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.324295 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.324305 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.324321 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.324333 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:27Z","lastTransitionTime":"2025-11-24T17:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.427835 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.427880 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.427891 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.427908 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.427927 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:27Z","lastTransitionTime":"2025-11-24T17:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.531919 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.531973 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.531982 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.532000 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.532010 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:27Z","lastTransitionTime":"2025-11-24T17:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.635986 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.636123 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.636144 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.636174 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.636194 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:27Z","lastTransitionTime":"2025-11-24T17:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.738691 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.738752 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.738828 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.738844 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.738855 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:27Z","lastTransitionTime":"2025-11-24T17:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.842011 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.842074 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.842087 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.842107 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.842119 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:27Z","lastTransitionTime":"2025-11-24T17:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.945313 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.945599 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.945697 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.945773 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:27 crc kubenswrapper[4808]: I1124 17:27:27.945867 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:27Z","lastTransitionTime":"2025-11-24T17:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.052908 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.052956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.052968 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.052984 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.052995 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:28Z","lastTransitionTime":"2025-11-24T17:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.155892 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.155930 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.155939 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.155954 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.155967 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:28Z","lastTransitionTime":"2025-11-24T17:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.258823 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.258899 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.258910 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.258928 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.258939 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:28Z","lastTransitionTime":"2025-11-24T17:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.346839 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.347122 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.347190 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:28 crc kubenswrapper[4808]: E1124 17:27:28.347127 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.347217 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:28 crc kubenswrapper[4808]: E1124 17:27:28.347405 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:28 crc kubenswrapper[4808]: E1124 17:27:28.347515 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:28 crc kubenswrapper[4808]: E1124 17:27:28.347624 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.364231 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.364263 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.364272 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.364289 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.364299 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:28Z","lastTransitionTime":"2025-11-24T17:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.370372 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:19Z\\\",\\\"message\\\":\\\" after 0 failed attempt(s)\\\\nI1124 17:27:19.478433 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1124 17:27:19.478436 6277 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1124 17:27:19.478438 6277 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:19.478416 6277 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1124 17:27:19.478439 6277 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1124 17:27:19.478380 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-q5m49 after 0 failed attempt(s)\\\\nI1124 17:27:19.478484 6277 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-q5m49\\\\nF1124 17:27:19.478364 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.386489 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.401967 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.420254 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.439271 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.454139 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.466707 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.466970 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.467133 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.467240 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.467320 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:28Z","lastTransitionTime":"2025-11-24T17:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.473140 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.483401 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.497383 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.518962 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.537320 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.551896 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.565338 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.569389 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.569436 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.569446 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.569467 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.569476 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:28Z","lastTransitionTime":"2025-11-24T17:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.580881 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.594745 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.608857 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.619848 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:28Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.672170 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.672215 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.672225 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.672243 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.672253 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:28Z","lastTransitionTime":"2025-11-24T17:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.775256 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.775292 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.775302 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.775316 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.775326 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:28Z","lastTransitionTime":"2025-11-24T17:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.876981 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.877061 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.877079 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.877101 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.877117 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:28Z","lastTransitionTime":"2025-11-24T17:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.979822 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.979960 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.979970 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.979985 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:28 crc kubenswrapper[4808]: I1124 17:27:28.979994 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:28Z","lastTransitionTime":"2025-11-24T17:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.081956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.081993 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.082002 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.082045 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.082065 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:29Z","lastTransitionTime":"2025-11-24T17:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.184724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.184761 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.184773 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.184791 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.184803 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:29Z","lastTransitionTime":"2025-11-24T17:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.287011 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.287069 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.287079 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.287097 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.287109 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:29Z","lastTransitionTime":"2025-11-24T17:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.388964 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.389032 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.389047 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.389065 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.389075 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:29Z","lastTransitionTime":"2025-11-24T17:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.491254 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.491306 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.491316 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.491334 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.491348 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:29Z","lastTransitionTime":"2025-11-24T17:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.593337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.593384 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.593393 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.593408 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.593418 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:29Z","lastTransitionTime":"2025-11-24T17:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.695584 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.695625 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.695641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.695657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.695666 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:29Z","lastTransitionTime":"2025-11-24T17:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.797597 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.797660 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.797682 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.797707 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.797724 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:29Z","lastTransitionTime":"2025-11-24T17:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.900763 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.900802 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.900814 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.900830 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:29 crc kubenswrapper[4808]: I1124 17:27:29.900841 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:29Z","lastTransitionTime":"2025-11-24T17:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.002829 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.002867 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.002880 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.002898 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.002909 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:30Z","lastTransitionTime":"2025-11-24T17:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.092095 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs\") pod \"network-metrics-daemon-8v2q4\" (UID: \"c26458eb-7f6c-42cb-b6c5-90de6756a994\") " pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:30 crc kubenswrapper[4808]: E1124 17:27:30.092264 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:27:30 crc kubenswrapper[4808]: E1124 17:27:30.092338 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs podName:c26458eb-7f6c-42cb-b6c5-90de6756a994 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:38.09231657 +0000 UTC m=+50.689984372 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs") pod "network-metrics-daemon-8v2q4" (UID: "c26458eb-7f6c-42cb-b6c5-90de6756a994") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.105771 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.105811 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.105820 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.105837 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.105846 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:30Z","lastTransitionTime":"2025-11-24T17:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.207560 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.207590 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.207598 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.207611 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.207621 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:30Z","lastTransitionTime":"2025-11-24T17:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.310315 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.310355 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.310366 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.310383 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.310393 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:30Z","lastTransitionTime":"2025-11-24T17:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.347082 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.347152 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:30 crc kubenswrapper[4808]: E1124 17:27:30.347237 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.347338 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.347584 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:30 crc kubenswrapper[4808]: E1124 17:27:30.347577 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:30 crc kubenswrapper[4808]: E1124 17:27:30.347645 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:30 crc kubenswrapper[4808]: E1124 17:27:30.347733 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.412920 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.412963 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.412973 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.412987 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.412997 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:30Z","lastTransitionTime":"2025-11-24T17:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.515479 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.515514 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.515545 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.515563 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.515574 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:30Z","lastTransitionTime":"2025-11-24T17:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.618454 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.618501 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.618510 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.618525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.618536 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:30Z","lastTransitionTime":"2025-11-24T17:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.720901 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.720938 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.720948 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.720963 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.720975 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:30Z","lastTransitionTime":"2025-11-24T17:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.823826 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.823872 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.823882 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.823899 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.823908 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:30Z","lastTransitionTime":"2025-11-24T17:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.927287 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.927352 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.927365 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.927387 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:30 crc kubenswrapper[4808]: I1124 17:27:30.927402 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:30Z","lastTransitionTime":"2025-11-24T17:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.030836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.030894 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.030905 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.030924 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.030937 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:31Z","lastTransitionTime":"2025-11-24T17:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.134149 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.134232 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.134254 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.134299 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.134340 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:31Z","lastTransitionTime":"2025-11-24T17:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.240535 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.240596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.240606 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.240626 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.240642 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:31Z","lastTransitionTime":"2025-11-24T17:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.344232 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.344282 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.344297 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.344317 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.344331 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:31Z","lastTransitionTime":"2025-11-24T17:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.347526 4808 scope.go:117] "RemoveContainer" containerID="bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.446850 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.446906 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.446919 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.446939 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.446968 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:31Z","lastTransitionTime":"2025-11-24T17:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.550152 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.550212 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.550227 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.550250 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.550263 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:31Z","lastTransitionTime":"2025-11-24T17:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.653604 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.653641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.653651 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.653669 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.653680 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:31Z","lastTransitionTime":"2025-11-24T17:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.675636 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovnkube-controller/1.log" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.679085 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerStarted","Data":"a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d"} Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.679261 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.711250 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.731490 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.758285 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.758325 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.758336 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.758349 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.758360 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:31Z","lastTransitionTime":"2025-11-24T17:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.759245 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.775793 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.793891 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.824308 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.839272 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.851100 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.860969 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.861004 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.861029 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.861052 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.861062 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:31Z","lastTransitionTime":"2025-11-24T17:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.863627 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.878226 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.893849 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.906154 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.922613 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.947315 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:19Z\\\",\\\"message\\\":\\\" after 0 failed attempt(s)\\\\nI1124 17:27:19.478433 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1124 17:27:19.478436 6277 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1124 17:27:19.478438 6277 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:19.478416 6277 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1124 17:27:19.478439 6277 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1124 17:27:19.478380 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-q5m49 after 0 failed attempt(s)\\\\nI1124 17:27:19.478484 6277 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-q5m49\\\\nF1124 17:27:19.478364 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.962704 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.963287 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.963337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.963377 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.963401 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.963414 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:31Z","lastTransitionTime":"2025-11-24T17:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.978834 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:31 crc kubenswrapper[4808]: I1124 17:27:31.993986 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:31Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.066228 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.066294 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.066307 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.066332 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.066347 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:32Z","lastTransitionTime":"2025-11-24T17:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.168893 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.168951 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.168962 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.168978 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.168987 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:32Z","lastTransitionTime":"2025-11-24T17:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.271322 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.271390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.271419 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.271447 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.271464 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:32Z","lastTransitionTime":"2025-11-24T17:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.347446 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.347498 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.347590 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:32 crc kubenswrapper[4808]: E1124 17:27:32.347641 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.347729 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:32 crc kubenswrapper[4808]: E1124 17:27:32.347970 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:32 crc kubenswrapper[4808]: E1124 17:27:32.348130 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:32 crc kubenswrapper[4808]: E1124 17:27:32.348288 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.373981 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.374057 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.374079 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.374106 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.374123 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:32Z","lastTransitionTime":"2025-11-24T17:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.477414 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.477487 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.477503 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.477520 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.477531 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:32Z","lastTransitionTime":"2025-11-24T17:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.586435 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.586493 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.586504 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.586523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.586536 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:32Z","lastTransitionTime":"2025-11-24T17:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.684639 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovnkube-controller/2.log" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.685379 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovnkube-controller/1.log" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.688674 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.688868 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.688886 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.688914 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.688926 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:32Z","lastTransitionTime":"2025-11-24T17:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.689003 4808 generic.go:334] "Generic (PLEG): container finished" podID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerID="a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d" exitCode=1 Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.689088 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerDied","Data":"a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d"} Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.689133 4808 scope.go:117] "RemoveContainer" containerID="bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.689865 4808 scope.go:117] "RemoveContainer" containerID="a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d" Nov 24 17:27:32 crc kubenswrapper[4808]: E1124 17:27:32.690042 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.709676 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.724683 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.745566 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:19Z\\\",\\\"message\\\":\\\" after 0 failed attempt(s)\\\\nI1124 17:27:19.478433 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1124 17:27:19.478436 6277 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1124 17:27:19.478438 6277 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:19.478416 6277 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1124 17:27:19.478439 6277 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1124 17:27:19.478380 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-q5m49 after 0 failed attempt(s)\\\\nI1124 17:27:19.478484 6277 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-q5m49\\\\nF1124 17:27:19.478364 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:32Z\\\",\\\"message\\\":\\\"-11-24T17:27:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 17:27:32.180486 6515 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180485 6515 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.139\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:17698, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 17:27:32.180494 6515 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180498 6515 services_controller.go:444] Built service openshift-apiserver/check-endpoints LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180509 6515 services_controller.go:445] Built service openshift-apiserver/check-endpoints LB template configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180523 6515 services_controller.go:451] Built service openshift-apiserver/check-endpoints cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protoco\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.761327 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.781702 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.791353 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.791398 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.791411 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.791430 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.791444 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:32Z","lastTransitionTime":"2025-11-24T17:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.797430 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.810319 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.827683 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.843414 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.855540 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.869155 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.889628 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.894384 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.894429 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.894442 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.894459 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.894471 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:32Z","lastTransitionTime":"2025-11-24T17:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.902194 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.913253 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.926005 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.940417 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.954085 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:32Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.996821 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.996858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.996870 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.996891 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:32 crc kubenswrapper[4808]: I1124 17:27:32.996906 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:32Z","lastTransitionTime":"2025-11-24T17:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.100066 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.100123 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.100142 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.100163 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.100180 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:33Z","lastTransitionTime":"2025-11-24T17:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.202882 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.202957 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.202972 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.202993 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.203007 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:33Z","lastTransitionTime":"2025-11-24T17:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.305101 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.305154 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.305163 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.305179 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.305188 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:33Z","lastTransitionTime":"2025-11-24T17:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.407811 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.408882 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.409184 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.409411 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.409620 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:33Z","lastTransitionTime":"2025-11-24T17:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.512472 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.512935 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.513032 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.513161 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.513243 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:33Z","lastTransitionTime":"2025-11-24T17:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.617758 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.617848 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.617869 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.618637 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.618711 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:33Z","lastTransitionTime":"2025-11-24T17:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.695733 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovnkube-controller/2.log" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.722373 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.722421 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.722434 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.722457 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.722471 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:33Z","lastTransitionTime":"2025-11-24T17:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.826002 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.826085 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.826101 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.826123 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.826138 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:33Z","lastTransitionTime":"2025-11-24T17:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.929937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.930373 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.930430 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.930457 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:33 crc kubenswrapper[4808]: I1124 17:27:33.930473 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:33Z","lastTransitionTime":"2025-11-24T17:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.033647 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.033704 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.033722 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.033750 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.033771 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:34Z","lastTransitionTime":"2025-11-24T17:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.136676 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.136726 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.136737 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.136754 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.136764 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:34Z","lastTransitionTime":"2025-11-24T17:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.244356 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.244435 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.244450 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.244472 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.244486 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:34Z","lastTransitionTime":"2025-11-24T17:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.346744 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.346763 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:34 crc kubenswrapper[4808]: E1124 17:27:34.346883 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.346942 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:34 crc kubenswrapper[4808]: E1124 17:27:34.347088 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.347178 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:34 crc kubenswrapper[4808]: E1124 17:27:34.347269 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:34 crc kubenswrapper[4808]: E1124 17:27:34.347353 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.347898 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.347971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.347985 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.348007 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.348049 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:34Z","lastTransitionTime":"2025-11-24T17:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.451002 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.451089 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.451101 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.451119 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.451130 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:34Z","lastTransitionTime":"2025-11-24T17:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.554724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.555122 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.555257 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.555365 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.555464 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:34Z","lastTransitionTime":"2025-11-24T17:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.658657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.658965 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.659076 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.659167 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.659263 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:34Z","lastTransitionTime":"2025-11-24T17:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.762321 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.762658 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.762736 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.762823 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.762920 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:34Z","lastTransitionTime":"2025-11-24T17:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.795350 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.795568 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.795631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.795692 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.795746 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:34Z","lastTransitionTime":"2025-11-24T17:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:34 crc kubenswrapper[4808]: E1124 17:27:34.808154 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:34Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.812472 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.812514 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.812525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.812542 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.812554 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:34Z","lastTransitionTime":"2025-11-24T17:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:34 crc kubenswrapper[4808]: E1124 17:27:34.823838 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:34Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.828949 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.829100 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.829188 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.829302 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.829433 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:34Z","lastTransitionTime":"2025-11-24T17:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:34 crc kubenswrapper[4808]: E1124 17:27:34.842737 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:34Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.849001 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.849054 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.849072 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.849090 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.849105 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:34Z","lastTransitionTime":"2025-11-24T17:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:34 crc kubenswrapper[4808]: E1124 17:27:34.864722 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:34Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.868859 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.868899 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.868912 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.868928 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.868938 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:34Z","lastTransitionTime":"2025-11-24T17:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:34 crc kubenswrapper[4808]: E1124 17:27:34.880497 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:34Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:34 crc kubenswrapper[4808]: E1124 17:27:34.880630 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.882798 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.882876 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.882893 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.882918 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.882933 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:34Z","lastTransitionTime":"2025-11-24T17:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.986232 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.986291 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.986307 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.986329 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:34 crc kubenswrapper[4808]: I1124 17:27:34.986345 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:34Z","lastTransitionTime":"2025-11-24T17:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.089724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.090177 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.090279 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.090382 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.090484 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:35Z","lastTransitionTime":"2025-11-24T17:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.193013 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.193120 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.193138 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.193163 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.193181 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:35Z","lastTransitionTime":"2025-11-24T17:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.295530 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.295805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.295867 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.295936 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.296039 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:35Z","lastTransitionTime":"2025-11-24T17:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.400291 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.400644 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.400721 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.400794 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.400862 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:35Z","lastTransitionTime":"2025-11-24T17:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.504995 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.505683 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.505776 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.505956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.506105 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:35Z","lastTransitionTime":"2025-11-24T17:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.609128 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.609191 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.609211 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.609241 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.609260 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:35Z","lastTransitionTime":"2025-11-24T17:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.712374 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.712435 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.712448 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.712470 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.712487 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:35Z","lastTransitionTime":"2025-11-24T17:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.820346 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.821291 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.821321 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.821354 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.821372 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:35Z","lastTransitionTime":"2025-11-24T17:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.924850 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.924903 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.924917 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.924937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:35 crc kubenswrapper[4808]: I1124 17:27:35.924951 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:35Z","lastTransitionTime":"2025-11-24T17:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.027870 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.027940 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.027951 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.027971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.027984 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:36Z","lastTransitionTime":"2025-11-24T17:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.131947 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.132038 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.132055 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.132078 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.132094 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:36Z","lastTransitionTime":"2025-11-24T17:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.234524 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.234579 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.234596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.234618 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.234635 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:36Z","lastTransitionTime":"2025-11-24T17:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.337681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.337772 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.337789 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.337814 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.337830 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:36Z","lastTransitionTime":"2025-11-24T17:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.347226 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.347280 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:36 crc kubenswrapper[4808]: E1124 17:27:36.347363 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.347431 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.347450 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:36 crc kubenswrapper[4808]: E1124 17:27:36.347718 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:36 crc kubenswrapper[4808]: E1124 17:27:36.347811 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:36 crc kubenswrapper[4808]: E1124 17:27:36.347924 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.440924 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.440971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.440983 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.440999 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.441010 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:36Z","lastTransitionTime":"2025-11-24T17:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.544845 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.544909 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.544920 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.544939 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.544968 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:36Z","lastTransitionTime":"2025-11-24T17:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.647846 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.647898 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.647907 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.647927 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.647941 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:36Z","lastTransitionTime":"2025-11-24T17:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.750943 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.750982 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.750996 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.751043 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.751055 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:36Z","lastTransitionTime":"2025-11-24T17:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.856465 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.856558 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.856583 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.857097 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.857376 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:36Z","lastTransitionTime":"2025-11-24T17:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.961106 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.961170 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.961190 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.961228 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:36 crc kubenswrapper[4808]: I1124 17:27:36.961248 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:36Z","lastTransitionTime":"2025-11-24T17:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.065094 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.065228 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.065245 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.065270 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.065285 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:37Z","lastTransitionTime":"2025-11-24T17:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.168742 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.168817 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.168826 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.168851 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.168872 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:37Z","lastTransitionTime":"2025-11-24T17:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.271242 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.271301 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.271317 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.271339 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.271356 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:37Z","lastTransitionTime":"2025-11-24T17:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.348729 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.367112 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.374037 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.374097 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.374118 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.374157 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.374177 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:37Z","lastTransitionTime":"2025-11-24T17:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.386203 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.422073 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.443549 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.460762 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.478571 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.478634 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.478644 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.478663 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.478674 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:37Z","lastTransitionTime":"2025-11-24T17:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.479214 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.501137 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.520655 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.533451 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.544544 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.558610 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.574380 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.581648 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.581692 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.581701 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.581720 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.581732 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:37Z","lastTransitionTime":"2025-11-24T17:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.592619 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.608768 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.625883 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.663596 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:19Z\\\",\\\"message\\\":\\\" after 0 failed attempt(s)\\\\nI1124 17:27:19.478433 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1124 17:27:19.478436 6277 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1124 17:27:19.478438 6277 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:19.478416 6277 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1124 17:27:19.478439 6277 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1124 17:27:19.478380 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-q5m49 after 0 failed attempt(s)\\\\nI1124 17:27:19.478484 6277 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-q5m49\\\\nF1124 17:27:19.478364 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:32Z\\\",\\\"message\\\":\\\"-11-24T17:27:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 17:27:32.180486 6515 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180485 6515 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.139\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:17698, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 17:27:32.180494 6515 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180498 6515 services_controller.go:444] Built service openshift-apiserver/check-endpoints LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180509 6515 services_controller.go:445] Built service openshift-apiserver/check-endpoints LB template configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180523 6515 services_controller.go:451] Built service openshift-apiserver/check-endpoints cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protoco\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.676991 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:37Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.683642 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.683677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.683687 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.683708 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.683981 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:37Z","lastTransitionTime":"2025-11-24T17:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.787364 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.787469 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.787487 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.787511 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.787527 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:37Z","lastTransitionTime":"2025-11-24T17:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.891730 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.891810 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.892204 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.892271 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.892322 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:37Z","lastTransitionTime":"2025-11-24T17:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.995286 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.995348 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.995362 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.995385 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:37 crc kubenswrapper[4808]: I1124 17:27:37.995399 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:37Z","lastTransitionTime":"2025-11-24T17:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.098913 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.098960 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.098970 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.098989 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.098999 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:38Z","lastTransitionTime":"2025-11-24T17:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.168790 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs\") pod \"network-metrics-daemon-8v2q4\" (UID: \"c26458eb-7f6c-42cb-b6c5-90de6756a994\") " pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:38 crc kubenswrapper[4808]: E1124 17:27:38.169111 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:27:38 crc kubenswrapper[4808]: E1124 17:27:38.169222 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs podName:c26458eb-7f6c-42cb-b6c5-90de6756a994 nodeName:}" failed. No retries permitted until 2025-11-24 17:27:54.169188122 +0000 UTC m=+66.766855954 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs") pod "network-metrics-daemon-8v2q4" (UID: "c26458eb-7f6c-42cb-b6c5-90de6756a994") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.203676 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.203752 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.203765 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.203794 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.203813 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:38Z","lastTransitionTime":"2025-11-24T17:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.317482 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.317542 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.317556 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.317580 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.317596 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:38Z","lastTransitionTime":"2025-11-24T17:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.347008 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.347074 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:38 crc kubenswrapper[4808]: E1124 17:27:38.347262 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.347089 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.347087 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:38 crc kubenswrapper[4808]: E1124 17:27:38.347395 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:38 crc kubenswrapper[4808]: E1124 17:27:38.347518 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:38 crc kubenswrapper[4808]: E1124 17:27:38.347598 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.363158 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.380973 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.401235 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.416074 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.420832 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.421025 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.421119 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.421233 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.421344 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:38Z","lastTransitionTime":"2025-11-24T17:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.434335 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.450987 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.466853 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.479206 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.489490 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.500624 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.511524 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.522518 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.524458 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.524525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.524540 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.524565 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.524580 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:38Z","lastTransitionTime":"2025-11-24T17:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.535276 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.547971 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.559232 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.577441 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf5c42490cf540701a5ca0fbd472abd5a1d8b7e6c0241b140f7793a2af0991c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:19Z\\\",\\\"message\\\":\\\" after 0 failed attempt(s)\\\\nI1124 17:27:19.478433 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1124 17:27:19.478436 6277 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1124 17:27:19.478438 6277 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:19.478416 6277 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1124 17:27:19.478439 6277 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1124 17:27:19.478380 6277 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-q5m49 after 0 failed attempt(s)\\\\nI1124 17:27:19.478484 6277 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-q5m49\\\\nF1124 17:27:19.478364 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:32Z\\\",\\\"message\\\":\\\"-11-24T17:27:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 17:27:32.180486 6515 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180485 6515 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.139\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:17698, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 17:27:32.180494 6515 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180498 6515 services_controller.go:444] Built service openshift-apiserver/check-endpoints LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180509 6515 services_controller.go:445] Built service openshift-apiserver/check-endpoints LB template configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180523 6515 services_controller.go:451] Built service openshift-apiserver/check-endpoints cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protoco\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.594107 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.603006 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.603746 4808 scope.go:117] "RemoveContainer" containerID="a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d" Nov 24 17:27:38 crc kubenswrapper[4808]: E1124 17:27:38.604126 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.615334 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.627272 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.627304 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.627314 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.627328 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.627337 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:38Z","lastTransitionTime":"2025-11-24T17:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.629413 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.640652 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.652703 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.671287 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.682860 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.691572 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.700696 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.717722 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.729764 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.729815 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.729828 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.729847 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.729860 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:38Z","lastTransitionTime":"2025-11-24T17:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.731341 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.742843 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.753657 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.773580 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:32Z\\\",\\\"message\\\":\\\"-11-24T17:27:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 17:27:32.180486 6515 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180485 6515 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.139\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:17698, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 17:27:32.180494 6515 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180498 6515 services_controller.go:444] Built service openshift-apiserver/check-endpoints LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180509 6515 services_controller.go:445] Built service openshift-apiserver/check-endpoints LB template configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180523 6515 services_controller.go:451] Built service openshift-apiserver/check-endpoints cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protoco\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.786553 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.799438 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.811835 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.823428 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:38Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.832301 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.832361 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.832374 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.832391 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.832403 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:38Z","lastTransitionTime":"2025-11-24T17:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.934775 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.934828 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.934842 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.934858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:38 crc kubenswrapper[4808]: I1124 17:27:38.935224 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:38Z","lastTransitionTime":"2025-11-24T17:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.037301 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.037349 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.037362 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.037382 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.037396 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:39Z","lastTransitionTime":"2025-11-24T17:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.140772 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.140819 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.140836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.140860 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.140876 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:39Z","lastTransitionTime":"2025-11-24T17:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.243273 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.243344 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.243360 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.243375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.243416 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:39Z","lastTransitionTime":"2025-11-24T17:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.347176 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.347224 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.347238 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.347259 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.347274 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:39Z","lastTransitionTime":"2025-11-24T17:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.450115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.450163 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.450179 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.450198 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.450210 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:39Z","lastTransitionTime":"2025-11-24T17:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.553325 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.553366 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.553377 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.553392 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.553402 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:39Z","lastTransitionTime":"2025-11-24T17:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.655627 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.655655 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.655665 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.655677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.655687 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:39Z","lastTransitionTime":"2025-11-24T17:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.758447 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.758507 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.758523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.758543 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.758559 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:39Z","lastTransitionTime":"2025-11-24T17:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.861512 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.861560 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.861592 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.861608 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.861622 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:39Z","lastTransitionTime":"2025-11-24T17:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.964109 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.964135 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.964143 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.964158 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:39 crc kubenswrapper[4808]: I1124 17:27:39.964167 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:39Z","lastTransitionTime":"2025-11-24T17:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.066644 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.066683 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.066693 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.066709 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.066717 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:40Z","lastTransitionTime":"2025-11-24T17:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.086630 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.086891 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:28:12.086857179 +0000 UTC m=+84.684524981 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.169046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.169103 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.169116 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.169135 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.169146 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:40Z","lastTransitionTime":"2025-11-24T17:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.187974 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.188067 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.188101 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.188132 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.188204 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.188253 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:28:12.188237873 +0000 UTC m=+84.785905675 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.188261 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.188400 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.188451 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.188485 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.188501 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.188522 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:28:12.18849674 +0000 UTC m=+84.786164702 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.188408 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.188566 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:28:12.188541541 +0000 UTC m=+84.786209363 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.188585 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.188670 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:28:12.188640324 +0000 UTC m=+84.786308306 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.271807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.271840 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.271849 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.271862 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.271871 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:40Z","lastTransitionTime":"2025-11-24T17:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.346769 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.346785 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.346783 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.346843 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.346952 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.347070 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.347163 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:40 crc kubenswrapper[4808]: E1124 17:27:40.347243 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.374717 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.374771 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.374782 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.374801 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.374814 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:40Z","lastTransitionTime":"2025-11-24T17:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.477466 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.477641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.477651 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.477665 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.477678 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:40Z","lastTransitionTime":"2025-11-24T17:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.580399 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.580436 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.580445 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.580460 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.580470 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:40Z","lastTransitionTime":"2025-11-24T17:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.682385 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.682436 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.682447 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.682460 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.682469 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:40Z","lastTransitionTime":"2025-11-24T17:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.753741 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.767715 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.768722 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:40Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.780138 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:40Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.784339 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.784383 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.784393 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.784409 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.784418 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:40Z","lastTransitionTime":"2025-11-24T17:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.792112 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:40Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.806442 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:40Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.818662 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:40Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.832594 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:40Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.868914 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:40Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.884925 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:40Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.889257 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.889314 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.889325 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.889361 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.889374 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:40Z","lastTransitionTime":"2025-11-24T17:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.898715 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:40Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.919983 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:40Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.935321 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:40Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.952655 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:40Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.969201 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:40Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.984480 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:40Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.992052 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.992121 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.992139 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.992164 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:40 crc kubenswrapper[4808]: I1124 17:27:40.992206 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:40Z","lastTransitionTime":"2025-11-24T17:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.006912 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:32Z\\\",\\\"message\\\":\\\"-11-24T17:27:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 17:27:32.180486 6515 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180485 6515 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.139\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:17698, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 17:27:32.180494 6515 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180498 6515 services_controller.go:444] Built service openshift-apiserver/check-endpoints LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180509 6515 services_controller.go:445] Built service openshift-apiserver/check-endpoints LB template configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180523 6515 services_controller.go:451] Built service openshift-apiserver/check-endpoints cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protoco\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:41Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.022596 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:41Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.038387 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:41Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.095400 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.095461 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.095471 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.095504 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.095516 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:41Z","lastTransitionTime":"2025-11-24T17:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.198859 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.198927 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.198946 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.198969 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.198986 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:41Z","lastTransitionTime":"2025-11-24T17:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.303036 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.303099 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.303113 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.303136 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.303151 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:41Z","lastTransitionTime":"2025-11-24T17:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.407343 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.407417 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.407433 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.407457 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.407473 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:41Z","lastTransitionTime":"2025-11-24T17:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.514478 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.514558 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.514574 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.514643 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.514667 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:41Z","lastTransitionTime":"2025-11-24T17:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.617574 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.617627 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.617639 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.617663 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.617678 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:41Z","lastTransitionTime":"2025-11-24T17:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.719929 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.720443 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.720456 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.720473 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.720486 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:41Z","lastTransitionTime":"2025-11-24T17:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.823069 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.823106 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.823115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.823135 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.823146 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:41Z","lastTransitionTime":"2025-11-24T17:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.925890 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.925937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.925949 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.925967 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:41 crc kubenswrapper[4808]: I1124 17:27:41.926289 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:41Z","lastTransitionTime":"2025-11-24T17:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.027752 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.027807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.027816 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.027831 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.027856 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:42Z","lastTransitionTime":"2025-11-24T17:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.130802 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.130847 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.130856 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.130871 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.130881 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:42Z","lastTransitionTime":"2025-11-24T17:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.233351 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.233385 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.233395 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.233408 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.233419 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:42Z","lastTransitionTime":"2025-11-24T17:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.336361 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.336431 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.336451 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.336480 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.336499 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:42Z","lastTransitionTime":"2025-11-24T17:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.346581 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.346636 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:42 crc kubenswrapper[4808]: E1124 17:27:42.346762 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.346827 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:42 crc kubenswrapper[4808]: E1124 17:27:42.346924 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:42 crc kubenswrapper[4808]: E1124 17:27:42.347079 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.347157 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:42 crc kubenswrapper[4808]: E1124 17:27:42.347248 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.439392 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.439438 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.439453 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.439472 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.439482 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:42Z","lastTransitionTime":"2025-11-24T17:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.541935 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.541973 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.541990 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.542044 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.542058 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:42Z","lastTransitionTime":"2025-11-24T17:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.644990 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.645262 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.645276 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.645293 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.645305 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:42Z","lastTransitionTime":"2025-11-24T17:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.747246 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.747294 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.747303 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.747315 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.747324 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:42Z","lastTransitionTime":"2025-11-24T17:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.849605 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.849655 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.849665 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.849680 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.849690 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:42Z","lastTransitionTime":"2025-11-24T17:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.951751 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.951817 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.951833 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.951855 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:42 crc kubenswrapper[4808]: I1124 17:27:42.951869 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:42Z","lastTransitionTime":"2025-11-24T17:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.054316 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.054360 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.054372 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.054390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.054403 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:43Z","lastTransitionTime":"2025-11-24T17:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.157247 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.157295 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.157311 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.157337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.157353 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:43Z","lastTransitionTime":"2025-11-24T17:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.259731 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.259772 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.259783 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.259800 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.259828 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:43Z","lastTransitionTime":"2025-11-24T17:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.362118 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.362169 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.362179 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.362195 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.362204 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:43Z","lastTransitionTime":"2025-11-24T17:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.465062 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.465110 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.465123 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.465142 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.465154 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:43Z","lastTransitionTime":"2025-11-24T17:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.567932 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.567972 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.567980 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.567996 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.568005 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:43Z","lastTransitionTime":"2025-11-24T17:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.672320 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.672375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.672390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.672410 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.672431 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:43Z","lastTransitionTime":"2025-11-24T17:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.778347 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.778386 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.778395 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.778410 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.778420 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:43Z","lastTransitionTime":"2025-11-24T17:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.881391 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.881460 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.881478 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.881511 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.881532 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:43Z","lastTransitionTime":"2025-11-24T17:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.983945 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.984007 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.984034 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.984053 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:43 crc kubenswrapper[4808]: I1124 17:27:43.984065 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:43Z","lastTransitionTime":"2025-11-24T17:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.086681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.086727 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.086740 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.086760 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.086772 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:44Z","lastTransitionTime":"2025-11-24T17:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.190245 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.190305 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.190315 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.190328 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.190339 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:44Z","lastTransitionTime":"2025-11-24T17:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.293206 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.293242 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.293252 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.293266 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.293277 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:44Z","lastTransitionTime":"2025-11-24T17:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.346501 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.346588 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:44 crc kubenswrapper[4808]: E1124 17:27:44.346635 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.346664 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:44 crc kubenswrapper[4808]: E1124 17:27:44.346797 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.346822 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:44 crc kubenswrapper[4808]: E1124 17:27:44.346892 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:44 crc kubenswrapper[4808]: E1124 17:27:44.346953 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.395781 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.395854 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.395878 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.395905 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.395925 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:44Z","lastTransitionTime":"2025-11-24T17:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.499688 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.499808 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.499833 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.499907 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.499933 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:44Z","lastTransitionTime":"2025-11-24T17:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.602737 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.602781 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.602791 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.602806 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.602816 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:44Z","lastTransitionTime":"2025-11-24T17:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.706557 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.706957 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.707227 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.707453 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.707680 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:44Z","lastTransitionTime":"2025-11-24T17:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.811466 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.812069 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.812172 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.812264 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.812345 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:44Z","lastTransitionTime":"2025-11-24T17:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.915585 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.915640 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.915653 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.915683 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:44 crc kubenswrapper[4808]: I1124 17:27:44.915699 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:44Z","lastTransitionTime":"2025-11-24T17:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.018558 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.018608 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.018621 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.018639 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.018652 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:45Z","lastTransitionTime":"2025-11-24T17:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.122242 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.122331 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.122354 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.122387 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.122408 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:45Z","lastTransitionTime":"2025-11-24T17:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.171050 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.171151 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.171176 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.171214 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.171241 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:45Z","lastTransitionTime":"2025-11-24T17:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:45 crc kubenswrapper[4808]: E1124 17:27:45.189806 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:45Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.196251 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.196330 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.196349 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.196378 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.196398 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:45Z","lastTransitionTime":"2025-11-24T17:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:45 crc kubenswrapper[4808]: E1124 17:27:45.215048 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:45Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.220928 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.221216 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.221329 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.221414 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.221480 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:45Z","lastTransitionTime":"2025-11-24T17:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:45 crc kubenswrapper[4808]: E1124 17:27:45.236136 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:45Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.240091 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.240285 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.240524 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.240631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.240710 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:45Z","lastTransitionTime":"2025-11-24T17:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:45 crc kubenswrapper[4808]: E1124 17:27:45.257950 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:45Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.263262 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.263311 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.263323 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.263352 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.263365 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:45Z","lastTransitionTime":"2025-11-24T17:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:45 crc kubenswrapper[4808]: E1124 17:27:45.277607 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:45Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:45 crc kubenswrapper[4808]: E1124 17:27:45.277876 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.280120 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.280289 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.280378 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.280460 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.280545 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:45Z","lastTransitionTime":"2025-11-24T17:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.383701 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.383777 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.383802 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.383837 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.383861 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:45Z","lastTransitionTime":"2025-11-24T17:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.486418 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.486451 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.486460 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.486474 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.486514 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:45Z","lastTransitionTime":"2025-11-24T17:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.589460 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.589487 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.589495 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.589509 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.589519 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:45Z","lastTransitionTime":"2025-11-24T17:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.692610 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.692658 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.692666 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.692682 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.692692 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:45Z","lastTransitionTime":"2025-11-24T17:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.796431 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.796497 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.796511 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.796536 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.796552 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:45Z","lastTransitionTime":"2025-11-24T17:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.900411 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.900456 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.900466 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.900517 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:45 crc kubenswrapper[4808]: I1124 17:27:45.900536 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:45Z","lastTransitionTime":"2025-11-24T17:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.002920 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.002982 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.002992 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.003008 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.003049 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:46Z","lastTransitionTime":"2025-11-24T17:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.105504 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.105546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.105557 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.105572 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.105584 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:46Z","lastTransitionTime":"2025-11-24T17:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.207685 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.207725 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.207761 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.207780 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.207792 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:46Z","lastTransitionTime":"2025-11-24T17:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.310589 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.310637 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.310649 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.310671 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.310683 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:46Z","lastTransitionTime":"2025-11-24T17:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.347332 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.347385 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:46 crc kubenswrapper[4808]: E1124 17:27:46.347469 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.347554 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:46 crc kubenswrapper[4808]: E1124 17:27:46.347605 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.347332 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:46 crc kubenswrapper[4808]: E1124 17:27:46.347856 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:46 crc kubenswrapper[4808]: E1124 17:27:46.347918 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.413186 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.413233 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.413248 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.413269 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.413283 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:46Z","lastTransitionTime":"2025-11-24T17:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.516546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.516593 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.516603 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.516619 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.516630 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:46Z","lastTransitionTime":"2025-11-24T17:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.619003 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.619053 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.619064 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.619080 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.619091 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:46Z","lastTransitionTime":"2025-11-24T17:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.721257 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.721300 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.721321 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.721339 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.721349 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:46Z","lastTransitionTime":"2025-11-24T17:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.823815 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.823857 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.823866 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.823883 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.823893 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:46Z","lastTransitionTime":"2025-11-24T17:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.926534 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.926566 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.926575 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.926590 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:46 crc kubenswrapper[4808]: I1124 17:27:46.926598 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:46Z","lastTransitionTime":"2025-11-24T17:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.029247 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.029295 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.029303 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.029320 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.029335 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:47Z","lastTransitionTime":"2025-11-24T17:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.131923 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.131967 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.131979 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.132000 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.132011 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:47Z","lastTransitionTime":"2025-11-24T17:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.234259 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.234315 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.234334 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.234351 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.234362 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:47Z","lastTransitionTime":"2025-11-24T17:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.337365 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.337451 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.337474 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.337505 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.337523 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:47Z","lastTransitionTime":"2025-11-24T17:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.440091 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.440133 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.440147 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.440167 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.440179 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:47Z","lastTransitionTime":"2025-11-24T17:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.542835 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.542874 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.542886 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.542903 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.542914 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:47Z","lastTransitionTime":"2025-11-24T17:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.646160 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.646226 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.646247 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.646278 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.646300 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:47Z","lastTransitionTime":"2025-11-24T17:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.749233 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.749283 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.749295 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.749320 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.749332 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:47Z","lastTransitionTime":"2025-11-24T17:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.852069 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.852118 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.852127 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.852147 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.852156 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:47Z","lastTransitionTime":"2025-11-24T17:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.956136 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.956180 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.956192 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.956209 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:47 crc kubenswrapper[4808]: I1124 17:27:47.956220 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:47Z","lastTransitionTime":"2025-11-24T17:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.060055 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.060106 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.060116 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.060136 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.060149 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:48Z","lastTransitionTime":"2025-11-24T17:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.161995 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.162046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.162055 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.162069 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.162078 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:48Z","lastTransitionTime":"2025-11-24T17:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.263733 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.263776 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.263786 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.263800 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.263812 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:48Z","lastTransitionTime":"2025-11-24T17:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.346450 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:48 crc kubenswrapper[4808]: E1124 17:27:48.346575 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.346605 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.346448 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:48 crc kubenswrapper[4808]: E1124 17:27:48.346740 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:48 crc kubenswrapper[4808]: E1124 17:27:48.346830 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.347149 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:48 crc kubenswrapper[4808]: E1124 17:27:48.347223 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.359128 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.366453 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.366490 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.366500 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.366515 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.366526 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:48Z","lastTransitionTime":"2025-11-24T17:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.369582 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.379141 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.390944 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.403588 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.415750 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.431959 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.456662 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:32Z\\\",\\\"message\\\":\\\"-11-24T17:27:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 17:27:32.180486 6515 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180485 6515 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.139\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:17698, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 17:27:32.180494 6515 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180498 6515 services_controller.go:444] Built service openshift-apiserver/check-endpoints LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180509 6515 services_controller.go:445] Built service openshift-apiserver/check-endpoints LB template configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180523 6515 services_controller.go:451] Built service openshift-apiserver/check-endpoints cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protoco\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.469368 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.469409 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.469419 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.469432 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.469443 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:48Z","lastTransitionTime":"2025-11-24T17:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.473183 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.489656 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"668040c0-c989-459b-9923-2f5a39a83958\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3462fe4e148a5c80763c265a819dcb04af31902463e9f1e34afefce17b13bbb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23297e0c2ec1bd342e97252d23b64dcf6039bcdb46579119053d8c654f953a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd30bb5725168f10a8a593c13456a101218f0bc17c492b7100470dee10fcf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.502045 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.515791 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.529508 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.542174 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.557182 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.569119 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.572053 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.572104 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.572118 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.572135 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.572145 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:48Z","lastTransitionTime":"2025-11-24T17:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.581266 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.600579 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:48Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.674352 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.674397 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.674409 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.674424 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.674437 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:48Z","lastTransitionTime":"2025-11-24T17:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.776799 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.776841 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.776851 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.776867 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.776877 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:48Z","lastTransitionTime":"2025-11-24T17:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.878794 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.878839 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.878850 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.878865 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.878879 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:48Z","lastTransitionTime":"2025-11-24T17:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.981483 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.981532 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.981545 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.981561 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:48 crc kubenswrapper[4808]: I1124 17:27:48.981571 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:48Z","lastTransitionTime":"2025-11-24T17:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.084633 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.084681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.084715 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.084734 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.084747 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:49Z","lastTransitionTime":"2025-11-24T17:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.188426 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.188465 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.188476 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.188492 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.188506 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:49Z","lastTransitionTime":"2025-11-24T17:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.292887 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.292925 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.292933 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.292947 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.292958 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:49Z","lastTransitionTime":"2025-11-24T17:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.348096 4808 scope.go:117] "RemoveContainer" containerID="a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d" Nov 24 17:27:49 crc kubenswrapper[4808]: E1124 17:27:49.348525 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.396334 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.396390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.396409 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.396427 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.396469 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:49Z","lastTransitionTime":"2025-11-24T17:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.500135 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.500219 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.500234 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.500257 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.500276 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:49Z","lastTransitionTime":"2025-11-24T17:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.603582 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.603623 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.603635 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.603649 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.603658 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:49Z","lastTransitionTime":"2025-11-24T17:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.705965 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.706009 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.706035 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.706053 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.706064 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:49Z","lastTransitionTime":"2025-11-24T17:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.807973 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.808016 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.808048 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.808065 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.808079 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:49Z","lastTransitionTime":"2025-11-24T17:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.911496 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.911534 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.911545 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.911561 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:49 crc kubenswrapper[4808]: I1124 17:27:49.911573 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:49Z","lastTransitionTime":"2025-11-24T17:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.014025 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.014073 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.014081 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.014096 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.014107 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:50Z","lastTransitionTime":"2025-11-24T17:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.117373 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.117499 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.117533 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.117581 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.117616 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:50Z","lastTransitionTime":"2025-11-24T17:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.225331 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.225399 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.225412 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.225434 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.225447 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:50Z","lastTransitionTime":"2025-11-24T17:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.328544 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.328607 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.328626 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.328654 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.328675 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:50Z","lastTransitionTime":"2025-11-24T17:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.347463 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.347620 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.347463 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:50 crc kubenswrapper[4808]: E1124 17:27:50.347807 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.347467 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:50 crc kubenswrapper[4808]: E1124 17:27:50.347883 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:50 crc kubenswrapper[4808]: E1124 17:27:50.347897 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:50 crc kubenswrapper[4808]: E1124 17:27:50.347623 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.432363 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.432424 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.432438 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.432462 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.432479 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:50Z","lastTransitionTime":"2025-11-24T17:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.536679 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.536756 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.536775 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.536799 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.536815 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:50Z","lastTransitionTime":"2025-11-24T17:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.639990 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.640031 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.640065 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.640085 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.640098 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:50Z","lastTransitionTime":"2025-11-24T17:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.752281 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.752345 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.752359 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.752382 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.752396 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:50Z","lastTransitionTime":"2025-11-24T17:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.855116 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.855167 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.855179 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.855199 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.855211 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:50Z","lastTransitionTime":"2025-11-24T17:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.958934 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.959003 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.959044 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.959071 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:50 crc kubenswrapper[4808]: I1124 17:27:50.959087 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:50Z","lastTransitionTime":"2025-11-24T17:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.062230 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.062317 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.062346 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.062386 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.062416 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:51Z","lastTransitionTime":"2025-11-24T17:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.166312 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.166466 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.166492 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.166525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.166552 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:51Z","lastTransitionTime":"2025-11-24T17:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.269370 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.269417 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.269427 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.269445 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.269453 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:51Z","lastTransitionTime":"2025-11-24T17:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.372759 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.372857 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.372880 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.372910 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.372930 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:51Z","lastTransitionTime":"2025-11-24T17:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.475760 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.475819 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.475832 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.475854 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.475869 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:51Z","lastTransitionTime":"2025-11-24T17:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.578856 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.578903 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.578911 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.578932 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.578946 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:51Z","lastTransitionTime":"2025-11-24T17:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.681347 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.681390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.681400 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.681417 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.681428 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:51Z","lastTransitionTime":"2025-11-24T17:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.784451 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.784499 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.784512 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.784535 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.784549 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:51Z","lastTransitionTime":"2025-11-24T17:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.887659 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.887704 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.887718 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.887739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.887753 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:51Z","lastTransitionTime":"2025-11-24T17:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.990575 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.990660 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.990674 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.990698 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:51 crc kubenswrapper[4808]: I1124 17:27:51.990719 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:51Z","lastTransitionTime":"2025-11-24T17:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.093933 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.093992 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.094002 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.094069 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.094081 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:52Z","lastTransitionTime":"2025-11-24T17:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.196767 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.196833 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.196864 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.196885 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.196898 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:52Z","lastTransitionTime":"2025-11-24T17:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.299080 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.299133 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.299149 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.299171 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.299187 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:52Z","lastTransitionTime":"2025-11-24T17:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.346448 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.346523 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.346526 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.346487 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:52 crc kubenswrapper[4808]: E1124 17:27:52.346637 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:52 crc kubenswrapper[4808]: E1124 17:27:52.346744 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:52 crc kubenswrapper[4808]: E1124 17:27:52.346892 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:52 crc kubenswrapper[4808]: E1124 17:27:52.346984 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.401734 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.401799 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.401807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.401828 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.401838 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:52Z","lastTransitionTime":"2025-11-24T17:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.504427 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.504490 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.504504 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.504529 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.504545 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:52Z","lastTransitionTime":"2025-11-24T17:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.608335 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.608436 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.608455 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.608484 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.608510 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:52Z","lastTransitionTime":"2025-11-24T17:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.711203 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.711297 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.711320 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.711347 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.711364 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:52Z","lastTransitionTime":"2025-11-24T17:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.814448 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.814507 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.814523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.814546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.814561 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:52Z","lastTransitionTime":"2025-11-24T17:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.917938 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.918005 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.918023 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.918082 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:52 crc kubenswrapper[4808]: I1124 17:27:52.918103 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:52Z","lastTransitionTime":"2025-11-24T17:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.050602 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.050646 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.050656 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.050671 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.050682 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:53Z","lastTransitionTime":"2025-11-24T17:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.153053 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.153113 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.153126 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.153141 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.153166 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:53Z","lastTransitionTime":"2025-11-24T17:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.255881 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.255947 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.255959 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.255985 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.256000 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:53Z","lastTransitionTime":"2025-11-24T17:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.358076 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.358234 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.358248 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.358264 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.358274 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:53Z","lastTransitionTime":"2025-11-24T17:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.460558 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.460611 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.460623 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.460643 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.460657 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:53Z","lastTransitionTime":"2025-11-24T17:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.563599 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.563647 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.563657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.563673 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.563685 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:53Z","lastTransitionTime":"2025-11-24T17:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.666992 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.667060 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.667071 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.667085 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.667095 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:53Z","lastTransitionTime":"2025-11-24T17:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.769854 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.769889 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.769901 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.769917 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.769928 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:53Z","lastTransitionTime":"2025-11-24T17:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.872774 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.872814 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.872824 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.872840 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.872853 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:53Z","lastTransitionTime":"2025-11-24T17:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.976580 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.976642 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.976652 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.976673 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:53 crc kubenswrapper[4808]: I1124 17:27:53.976685 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:53Z","lastTransitionTime":"2025-11-24T17:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.079728 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.079795 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.079807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.079826 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.079842 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:54Z","lastTransitionTime":"2025-11-24T17:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.182693 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.182736 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.182747 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.182765 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.182775 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:54Z","lastTransitionTime":"2025-11-24T17:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.263017 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs\") pod \"network-metrics-daemon-8v2q4\" (UID: \"c26458eb-7f6c-42cb-b6c5-90de6756a994\") " pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:54 crc kubenswrapper[4808]: E1124 17:27:54.263232 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:27:54 crc kubenswrapper[4808]: E1124 17:27:54.263356 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs podName:c26458eb-7f6c-42cb-b6c5-90de6756a994 nodeName:}" failed. No retries permitted until 2025-11-24 17:28:26.263332378 +0000 UTC m=+98.861000180 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs") pod "network-metrics-daemon-8v2q4" (UID: "c26458eb-7f6c-42cb-b6c5-90de6756a994") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.285163 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.285234 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.285246 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.285266 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.285281 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:54Z","lastTransitionTime":"2025-11-24T17:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.346631 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.346720 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.346760 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.346646 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:54 crc kubenswrapper[4808]: E1124 17:27:54.346882 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:54 crc kubenswrapper[4808]: E1124 17:27:54.346963 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:54 crc kubenswrapper[4808]: E1124 17:27:54.347047 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:54 crc kubenswrapper[4808]: E1124 17:27:54.347116 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.386954 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.387006 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.387039 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.387056 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.387066 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:54Z","lastTransitionTime":"2025-11-24T17:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.489613 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.489680 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.489696 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.489718 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.489733 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:54Z","lastTransitionTime":"2025-11-24T17:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.592655 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.592704 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.592716 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.592735 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.592747 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:54Z","lastTransitionTime":"2025-11-24T17:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.694810 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.694851 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.694865 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.694880 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.694891 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:54Z","lastTransitionTime":"2025-11-24T17:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.796658 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.796696 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.796708 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.796722 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.796731 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:54Z","lastTransitionTime":"2025-11-24T17:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.899909 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.899960 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.899971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.899986 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:54 crc kubenswrapper[4808]: I1124 17:27:54.899997 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:54Z","lastTransitionTime":"2025-11-24T17:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.003156 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.003211 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.003224 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.003247 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.003268 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:55Z","lastTransitionTime":"2025-11-24T17:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.105688 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.105726 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.105743 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.105765 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.105779 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:55Z","lastTransitionTime":"2025-11-24T17:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.208151 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.208192 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.208203 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.208221 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.208233 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:55Z","lastTransitionTime":"2025-11-24T17:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.311047 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.311089 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.311102 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.311119 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.311131 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:55Z","lastTransitionTime":"2025-11-24T17:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.413555 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.413599 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.413611 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.413628 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.413639 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:55Z","lastTransitionTime":"2025-11-24T17:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.489309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.489354 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.489367 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.489385 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.489397 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:55Z","lastTransitionTime":"2025-11-24T17:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:55 crc kubenswrapper[4808]: E1124 17:27:55.505141 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:55Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.509900 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.509950 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.509963 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.509981 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.509995 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:55Z","lastTransitionTime":"2025-11-24T17:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:55 crc kubenswrapper[4808]: E1124 17:27:55.522496 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:55Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.526650 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.526691 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.526701 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.526717 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.526727 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:55Z","lastTransitionTime":"2025-11-24T17:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:55 crc kubenswrapper[4808]: E1124 17:27:55.539766 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:55Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.544444 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.544506 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.544521 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.544545 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.544560 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:55Z","lastTransitionTime":"2025-11-24T17:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:55 crc kubenswrapper[4808]: E1124 17:27:55.602303 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:55Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.606335 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.606355 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.606364 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.606377 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.606386 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:55Z","lastTransitionTime":"2025-11-24T17:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:55 crc kubenswrapper[4808]: E1124 17:27:55.619681 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:55Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:55 crc kubenswrapper[4808]: E1124 17:27:55.619797 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.621646 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.621670 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.621680 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.621694 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.621705 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:55Z","lastTransitionTime":"2025-11-24T17:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.724340 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.724515 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.724547 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.724578 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.724600 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:55Z","lastTransitionTime":"2025-11-24T17:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.827117 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.827194 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.827219 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.827249 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.827271 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:55Z","lastTransitionTime":"2025-11-24T17:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.929244 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.929287 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.929296 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.929310 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:55 crc kubenswrapper[4808]: I1124 17:27:55.929319 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:55Z","lastTransitionTime":"2025-11-24T17:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.032588 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.032657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.032676 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.032702 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.032720 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:56Z","lastTransitionTime":"2025-11-24T17:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.135408 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.135439 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.135448 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.135465 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.135478 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:56Z","lastTransitionTime":"2025-11-24T17:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.238507 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.238556 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.238569 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.238589 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.238603 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:56Z","lastTransitionTime":"2025-11-24T17:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.340978 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.341045 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.341054 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.341069 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.341078 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:56Z","lastTransitionTime":"2025-11-24T17:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.346620 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.346623 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.346756 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.346904 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:56 crc kubenswrapper[4808]: E1124 17:27:56.346932 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:56 crc kubenswrapper[4808]: E1124 17:27:56.347029 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:56 crc kubenswrapper[4808]: E1124 17:27:56.347107 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:56 crc kubenswrapper[4808]: E1124 17:27:56.347260 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.443705 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.443747 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.443759 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.443778 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.443792 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:56Z","lastTransitionTime":"2025-11-24T17:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.545675 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.545720 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.545730 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.545743 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.545751 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:56Z","lastTransitionTime":"2025-11-24T17:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.648062 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.648110 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.648123 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.648139 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.648150 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:56Z","lastTransitionTime":"2025-11-24T17:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.751072 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.751136 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.751148 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.751173 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.751187 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:56Z","lastTransitionTime":"2025-11-24T17:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.792735 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q5m49_6a192cd6-853d-4643-855d-ac61168b1cda/kube-multus/0.log" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.792796 4808 generic.go:334] "Generic (PLEG): container finished" podID="6a192cd6-853d-4643-855d-ac61168b1cda" containerID="e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147" exitCode=1 Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.792859 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q5m49" event={"ID":"6a192cd6-853d-4643-855d-ac61168b1cda","Type":"ContainerDied","Data":"e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147"} Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.794425 4808 scope.go:117] "RemoveContainer" containerID="e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.810699 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:56Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.826941 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:56Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.840276 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:56Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.856299 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.856354 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.856365 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.856385 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.856400 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:56Z","lastTransitionTime":"2025-11-24T17:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.856564 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:56Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.872623 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:56Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.890269 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:56Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.905156 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:56Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.916377 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:56Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.936332 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:32Z\\\",\\\"message\\\":\\\"-11-24T17:27:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 17:27:32.180486 6515 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180485 6515 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.139\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:17698, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 17:27:32.180494 6515 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180498 6515 services_controller.go:444] Built service openshift-apiserver/check-endpoints LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180509 6515 services_controller.go:445] Built service openshift-apiserver/check-endpoints LB template configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180523 6515 services_controller.go:451] Built service openshift-apiserver/check-endpoints cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protoco\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:56Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.950296 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"668040c0-c989-459b-9923-2f5a39a83958\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3462fe4e148a5c80763c265a819dcb04af31902463e9f1e34afefce17b13bbb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23297e0c2ec1bd342e97252d23b64dcf6039bcdb46579119053d8c654f953a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd30bb5725168f10a8a593c13456a101218f0bc17c492b7100470dee10fcf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:56Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.959312 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.959382 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.959396 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.959416 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.959429 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:56Z","lastTransitionTime":"2025-11-24T17:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.967017 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:56Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.981826 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:56Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:56 crc kubenswrapper[4808]: I1124 17:27:56.996677 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:56Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.019690 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.032532 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.046373 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.060484 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:56Z\\\",\\\"message\\\":\\\"2025-11-24T17:27:11+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b\\\\n2025-11-24T17:27:11+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b to /host/opt/cni/bin/\\\\n2025-11-24T17:27:11Z [verbose] multus-daemon started\\\\n2025-11-24T17:27:11Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:27:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.062549 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.062614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.062626 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.062647 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.062661 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:57Z","lastTransitionTime":"2025-11-24T17:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.078346 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.166127 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.166188 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.166203 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.166229 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.166244 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:57Z","lastTransitionTime":"2025-11-24T17:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.269298 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.269357 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.269369 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.269390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.269405 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:57Z","lastTransitionTime":"2025-11-24T17:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.372732 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.372783 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.372797 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.372816 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.372827 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:57Z","lastTransitionTime":"2025-11-24T17:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.475796 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.475860 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.475871 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.475888 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.475901 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:57Z","lastTransitionTime":"2025-11-24T17:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.579163 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.579220 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.579230 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.579247 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.579260 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:57Z","lastTransitionTime":"2025-11-24T17:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.681734 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.681781 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.681792 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.681812 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.681826 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:57Z","lastTransitionTime":"2025-11-24T17:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.784235 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.784289 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.784302 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.784323 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.784337 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:57Z","lastTransitionTime":"2025-11-24T17:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.798388 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q5m49_6a192cd6-853d-4643-855d-ac61168b1cda/kube-multus/0.log" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.798470 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q5m49" event={"ID":"6a192cd6-853d-4643-855d-ac61168b1cda","Type":"ContainerStarted","Data":"3ac714762891a8496f5f9fb3d368c22acffd162a1a646871d2be0386c78b0a21"} Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.812774 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.825680 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.846944 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.860529 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.872684 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.885278 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac714762891a8496f5f9fb3d368c22acffd162a1a646871d2be0386c78b0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:56Z\\\",\\\"message\\\":\\\"2025-11-24T17:27:11+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b\\\\n2025-11-24T17:27:11+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b to /host/opt/cni/bin/\\\\n2025-11-24T17:27:11Z [verbose] multus-daemon started\\\\n2025-11-24T17:27:11Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:27:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.886557 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.886604 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.886615 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.886631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.886640 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:57Z","lastTransitionTime":"2025-11-24T17:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.901557 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.914265 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.927951 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.941576 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.953283 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.965082 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.980711 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.989497 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.989545 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.989556 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.989576 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.989590 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:57Z","lastTransitionTime":"2025-11-24T17:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:57 crc kubenswrapper[4808]: I1124 17:27:57.993407 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:57Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.004591 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.022390 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:32Z\\\",\\\"message\\\":\\\"-11-24T17:27:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 17:27:32.180486 6515 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180485 6515 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.139\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:17698, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 17:27:32.180494 6515 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180498 6515 services_controller.go:444] Built service openshift-apiserver/check-endpoints LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180509 6515 services_controller.go:445] Built service openshift-apiserver/check-endpoints LB template configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180523 6515 services_controller.go:451] Built service openshift-apiserver/check-endpoints cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protoco\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.034350 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"668040c0-c989-459b-9923-2f5a39a83958\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3462fe4e148a5c80763c265a819dcb04af31902463e9f1e34afefce17b13bbb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23297e0c2ec1bd342e97252d23b64dcf6039bcdb46579119053d8c654f953a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd30bb5725168f10a8a593c13456a101218f0bc17c492b7100470dee10fcf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.047279 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.092711 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.092761 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.092777 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.092798 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.092812 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:58Z","lastTransitionTime":"2025-11-24T17:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.194958 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.195045 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.195057 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.195077 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.195094 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:58Z","lastTransitionTime":"2025-11-24T17:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.298155 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.298208 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.298221 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.298241 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.298256 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:58Z","lastTransitionTime":"2025-11-24T17:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.346730 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.346845 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.346858 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:27:58 crc kubenswrapper[4808]: E1124 17:27:58.347126 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:27:58 crc kubenswrapper[4808]: E1124 17:27:58.347205 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.347461 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:27:58 crc kubenswrapper[4808]: E1124 17:27:58.347537 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:27:58 crc kubenswrapper[4808]: E1124 17:27:58.347648 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.358306 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"668040c0-c989-459b-9923-2f5a39a83958\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3462fe4e148a5c80763c265a819dcb04af31902463e9f1e34afefce17b13bbb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23297e0c2ec1bd342e97252d23b64dcf6039bcdb46579119053d8c654f953a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd30bb5725168f10a8a593c13456a101218f0bc17c492b7100470dee10fcf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.373380 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.383875 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.398523 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.401163 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.401221 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.401232 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.401248 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.401259 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:58Z","lastTransitionTime":"2025-11-24T17:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.423105 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.436150 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.448011 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.459078 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac714762891a8496f5f9fb3d368c22acffd162a1a646871d2be0386c78b0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:56Z\\\",\\\"message\\\":\\\"2025-11-24T17:27:11+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b\\\\n2025-11-24T17:27:11+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b to /host/opt/cni/bin/\\\\n2025-11-24T17:27:11Z [verbose] multus-daemon started\\\\n2025-11-24T17:27:11Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:27:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.476358 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.488221 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.503258 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.504768 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.504827 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.504840 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.504865 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.504879 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:58Z","lastTransitionTime":"2025-11-24T17:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.516219 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.527473 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.539921 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.557561 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.572465 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.585519 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.607475 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.607525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.607534 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.607554 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.607569 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:58Z","lastTransitionTime":"2025-11-24T17:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.607606 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:32Z\\\",\\\"message\\\":\\\"-11-24T17:27:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 17:27:32.180486 6515 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180485 6515 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.139\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:17698, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 17:27:32.180494 6515 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180498 6515 services_controller.go:444] Built service openshift-apiserver/check-endpoints LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180509 6515 services_controller.go:445] Built service openshift-apiserver/check-endpoints LB template configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180523 6515 services_controller.go:451] Built service openshift-apiserver/check-endpoints cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protoco\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:27:58Z is after 2025-08-24T17:21:41Z" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.709931 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.709982 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.709993 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.710008 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.710033 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:58Z","lastTransitionTime":"2025-11-24T17:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.812176 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.812227 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.812239 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.812266 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.812283 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:58Z","lastTransitionTime":"2025-11-24T17:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.915569 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.915631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.915645 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.915670 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:58 crc kubenswrapper[4808]: I1124 17:27:58.915689 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:58Z","lastTransitionTime":"2025-11-24T17:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.018896 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.018943 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.018956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.018977 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.018989 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:59Z","lastTransitionTime":"2025-11-24T17:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.122163 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.122205 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.122214 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.122230 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.122242 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:59Z","lastTransitionTime":"2025-11-24T17:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.228224 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.228272 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.228282 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.228305 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.228317 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:59Z","lastTransitionTime":"2025-11-24T17:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.330885 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.330922 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.330931 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.330946 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.330955 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:59Z","lastTransitionTime":"2025-11-24T17:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.435088 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.435139 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.435152 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.435170 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.435180 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:59Z","lastTransitionTime":"2025-11-24T17:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.538295 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.538345 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.538355 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.538374 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.538386 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:59Z","lastTransitionTime":"2025-11-24T17:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.640822 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.641483 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.641519 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.641540 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.641557 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:59Z","lastTransitionTime":"2025-11-24T17:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.743366 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.743423 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.743437 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.743453 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.743464 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:59Z","lastTransitionTime":"2025-11-24T17:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.845182 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.845250 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.845259 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.845274 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.845284 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:59Z","lastTransitionTime":"2025-11-24T17:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.948115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.948152 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.948160 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.948175 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:27:59 crc kubenswrapper[4808]: I1124 17:27:59.948184 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:27:59Z","lastTransitionTime":"2025-11-24T17:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.051553 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.051622 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.051633 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.051658 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.051671 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:00Z","lastTransitionTime":"2025-11-24T17:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.153743 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.153802 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.153813 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.153838 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.153851 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:00Z","lastTransitionTime":"2025-11-24T17:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.256122 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.256175 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.256209 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.256227 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.256241 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:00Z","lastTransitionTime":"2025-11-24T17:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.347191 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.347262 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.347277 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.347191 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:00 crc kubenswrapper[4808]: E1124 17:28:00.347360 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:00 crc kubenswrapper[4808]: E1124 17:28:00.347438 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:00 crc kubenswrapper[4808]: E1124 17:28:00.347590 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:00 crc kubenswrapper[4808]: E1124 17:28:00.347688 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.358866 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.358917 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.358930 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.358949 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.358960 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:00Z","lastTransitionTime":"2025-11-24T17:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.461643 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.461760 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.461786 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.461848 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.461877 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:00Z","lastTransitionTime":"2025-11-24T17:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.565985 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.566081 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.566098 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.566120 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.566139 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:00Z","lastTransitionTime":"2025-11-24T17:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.668460 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.668509 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.668518 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.668534 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.668544 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:00Z","lastTransitionTime":"2025-11-24T17:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.770742 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.770817 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.770830 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.770853 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.770864 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:00Z","lastTransitionTime":"2025-11-24T17:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.873977 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.874038 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.874048 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.874066 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.874078 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:00Z","lastTransitionTime":"2025-11-24T17:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.982091 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.982579 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.982672 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.982791 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:00 crc kubenswrapper[4808]: I1124 17:28:00.982881 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:00Z","lastTransitionTime":"2025-11-24T17:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.086115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.086189 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.086202 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.086224 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.086237 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:01Z","lastTransitionTime":"2025-11-24T17:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.188595 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.188648 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.188660 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.188678 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.188692 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:01Z","lastTransitionTime":"2025-11-24T17:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.290897 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.290949 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.290960 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.290979 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.290995 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:01Z","lastTransitionTime":"2025-11-24T17:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.394110 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.394154 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.394167 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.394184 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.394194 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:01Z","lastTransitionTime":"2025-11-24T17:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.496461 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.496524 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.496535 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.496562 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.496577 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:01Z","lastTransitionTime":"2025-11-24T17:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.599266 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.599329 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.599341 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.599361 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.599374 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:01Z","lastTransitionTime":"2025-11-24T17:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.702438 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.702494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.702507 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.702528 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.702543 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:01Z","lastTransitionTime":"2025-11-24T17:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.808297 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.808353 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.808365 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.808383 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.808401 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:01Z","lastTransitionTime":"2025-11-24T17:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.911150 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.911228 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.911246 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.911274 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:01 crc kubenswrapper[4808]: I1124 17:28:01.911292 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:01Z","lastTransitionTime":"2025-11-24T17:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.014729 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.014792 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.014808 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.014829 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.014843 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:02Z","lastTransitionTime":"2025-11-24T17:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.117359 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.117405 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.117416 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.117434 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.117447 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:02Z","lastTransitionTime":"2025-11-24T17:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.220491 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.220573 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.220590 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.220616 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.220633 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:02Z","lastTransitionTime":"2025-11-24T17:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.323377 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.323421 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.323432 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.323447 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.323456 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:02Z","lastTransitionTime":"2025-11-24T17:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.346995 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.347112 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.347145 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.347668 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:02 crc kubenswrapper[4808]: E1124 17:28:02.347819 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:02 crc kubenswrapper[4808]: E1124 17:28:02.348039 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:02 crc kubenswrapper[4808]: E1124 17:28:02.348239 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:02 crc kubenswrapper[4808]: E1124 17:28:02.348349 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.427809 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.427873 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.427887 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.427915 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.427929 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:02Z","lastTransitionTime":"2025-11-24T17:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.530934 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.530993 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.531006 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.531048 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.531063 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:02Z","lastTransitionTime":"2025-11-24T17:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.634341 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.634413 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.634432 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.634464 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.634491 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:02Z","lastTransitionTime":"2025-11-24T17:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.737439 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.737521 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.737542 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.737570 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.737605 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:02Z","lastTransitionTime":"2025-11-24T17:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.840104 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.840154 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.840167 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.840187 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.840200 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:02Z","lastTransitionTime":"2025-11-24T17:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.942875 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.942929 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.942939 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.942961 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:02 crc kubenswrapper[4808]: I1124 17:28:02.942978 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:02Z","lastTransitionTime":"2025-11-24T17:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.052253 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.052304 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.052314 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.052330 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.052340 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:03Z","lastTransitionTime":"2025-11-24T17:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.155721 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.155776 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.155787 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.155807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.155820 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:03Z","lastTransitionTime":"2025-11-24T17:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.258559 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.258614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.258623 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.258644 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.258656 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:03Z","lastTransitionTime":"2025-11-24T17:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.347822 4808 scope.go:117] "RemoveContainer" containerID="a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.360745 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.360782 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.360792 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.360807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.360819 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:03Z","lastTransitionTime":"2025-11-24T17:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.462900 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.463305 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.463313 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.463327 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.463339 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:03Z","lastTransitionTime":"2025-11-24T17:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.566055 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.566088 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.566100 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.566117 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.566129 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:03Z","lastTransitionTime":"2025-11-24T17:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.668921 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.668984 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.668999 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.669060 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.669075 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:03Z","lastTransitionTime":"2025-11-24T17:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.771854 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.771898 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.771909 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.771929 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.771942 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:03Z","lastTransitionTime":"2025-11-24T17:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.819969 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovnkube-controller/2.log" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.822443 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerStarted","Data":"3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222"} Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.822988 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.842068 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.869658 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.874338 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.874401 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.874414 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.874441 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.874455 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:03Z","lastTransitionTime":"2025-11-24T17:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.886862 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.902228 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.912911 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.931596 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:32Z\\\",\\\"message\\\":\\\"-11-24T17:27:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 17:27:32.180486 6515 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180485 6515 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.139\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:17698, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 17:27:32.180494 6515 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180498 6515 services_controller.go:444] Built service openshift-apiserver/check-endpoints LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180509 6515 services_controller.go:445] Built service openshift-apiserver/check-endpoints LB template configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180523 6515 services_controller.go:451] Built service openshift-apiserver/check-endpoints cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protoco\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.946242 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.959012 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.971326 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.976811 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.976843 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.976856 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.976878 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.976892 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:03Z","lastTransitionTime":"2025-11-24T17:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:03 crc kubenswrapper[4808]: I1124 17:28:03.984972 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"668040c0-c989-459b-9923-2f5a39a83958\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3462fe4e148a5c80763c265a819dcb04af31902463e9f1e34afefce17b13bbb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23297e0c2ec1bd342e97252d23b64dcf6039bcdb46579119053d8c654f953a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd30bb5725168f10a8a593c13456a101218f0bc17c492b7100470dee10fcf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.000426 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.014460 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac714762891a8496f5f9fb3d368c22acffd162a1a646871d2be0386c78b0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:56Z\\\",\\\"message\\\":\\\"2025-11-24T17:27:11+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b\\\\n2025-11-24T17:27:11+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b to /host/opt/cni/bin/\\\\n2025-11-24T17:27:11Z [verbose] multus-daemon started\\\\n2025-11-24T17:27:11Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:27:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.029710 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.040919 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.057003 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.078163 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.079491 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.079525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.079536 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.079557 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.079571 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:04Z","lastTransitionTime":"2025-11-24T17:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.091556 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.102467 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.182907 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.182973 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.182985 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.183008 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.183039 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:04Z","lastTransitionTime":"2025-11-24T17:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.285571 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.285608 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.285621 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.285636 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.285647 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:04Z","lastTransitionTime":"2025-11-24T17:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.347446 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.347513 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:04 crc kubenswrapper[4808]: E1124 17:28:04.347621 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.347704 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.347447 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:04 crc kubenswrapper[4808]: E1124 17:28:04.347839 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:04 crc kubenswrapper[4808]: E1124 17:28:04.347878 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:04 crc kubenswrapper[4808]: E1124 17:28:04.347785 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.388187 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.388244 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.388257 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.388282 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.388297 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:04Z","lastTransitionTime":"2025-11-24T17:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.490874 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.490925 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.490940 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.490960 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.490975 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:04Z","lastTransitionTime":"2025-11-24T17:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.594613 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.594668 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.594682 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.594707 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.594725 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:04Z","lastTransitionTime":"2025-11-24T17:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.697876 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.697923 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.697936 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.697956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.697969 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:04Z","lastTransitionTime":"2025-11-24T17:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.801150 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.801226 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.801241 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.801261 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.801277 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:04Z","lastTransitionTime":"2025-11-24T17:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.828503 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovnkube-controller/3.log" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.829778 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovnkube-controller/2.log" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.832965 4808 generic.go:334] "Generic (PLEG): container finished" podID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerID="3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222" exitCode=1 Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.833066 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerDied","Data":"3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222"} Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.833135 4808 scope.go:117] "RemoveContainer" containerID="a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.833869 4808 scope.go:117] "RemoveContainer" containerID="3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222" Nov 24 17:28:04 crc kubenswrapper[4808]: E1124 17:28:04.834210 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.849804 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"668040c0-c989-459b-9923-2f5a39a83958\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3462fe4e148a5c80763c265a819dcb04af31902463e9f1e34afefce17b13bbb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23297e0c2ec1bd342e97252d23b64dcf6039bcdb46579119053d8c654f953a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd30bb5725168f10a8a593c13456a101218f0bc17c492b7100470dee10fcf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.866477 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.883373 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.895528 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.904264 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.904338 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.904352 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.904374 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.904390 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:04Z","lastTransitionTime":"2025-11-24T17:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.912015 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac714762891a8496f5f9fb3d368c22acffd162a1a646871d2be0386c78b0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:56Z\\\",\\\"message\\\":\\\"2025-11-24T17:27:11+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b\\\\n2025-11-24T17:27:11+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b to /host/opt/cni/bin/\\\\n2025-11-24T17:27:11Z [verbose] multus-daemon started\\\\n2025-11-24T17:27:11Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:27:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.929998 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.942470 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.955688 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.975775 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:04 crc kubenswrapper[4808]: I1124 17:28:04.988221 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.000834 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:04Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.006881 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.006928 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.006942 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.006963 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.006977 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:05Z","lastTransitionTime":"2025-11-24T17:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.013956 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.028952 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.042224 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.056507 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.071849 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.092934 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a433deb1ab18daba6226ce0e8106416d62e561dc281a58e32b441c774fc0750d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:32Z\\\",\\\"message\\\":\\\"-11-24T17:27:32Z is after 2025-08-24T17:21:41Z]\\\\nI1124 17:27:32.180486 6515 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180485 6515 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.139\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:17698, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 17:27:32.180494 6515 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:27:32.180498 6515 services_controller.go:444] Built service openshift-apiserver/check-endpoints LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180509 6515 services_controller.go:445] Built service openshift-apiserver/check-endpoints LB template configs for network=default: []services.lbConfig(nil)\\\\nI1124 17:27:32.180523 6515 services_controller.go:451] Built service openshift-apiserver/check-endpoints cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protoco\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:28:04Z\\\",\\\"message\\\":\\\"LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/catalog-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.204\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 17:28:04.116754 6882 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:28:04.116754 6882 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:28:04.116760 6882 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-ht87w in node crc\\\\nI1124 17:28:04.116764 6882 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:28:04.116770 6882 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-ht87w after 0 failed atte\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.107282 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.108950 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.108984 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.108999 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.109049 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.109063 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:05Z","lastTransitionTime":"2025-11-24T17:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.211952 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.211997 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.212010 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.212056 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.212069 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:05Z","lastTransitionTime":"2025-11-24T17:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.315176 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.315223 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.315235 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.315254 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.315266 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:05Z","lastTransitionTime":"2025-11-24T17:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.418589 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.418641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.418654 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.418675 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.418687 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:05Z","lastTransitionTime":"2025-11-24T17:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.521160 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.521246 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.521256 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.521275 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.521290 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:05Z","lastTransitionTime":"2025-11-24T17:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.624999 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.625095 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.625105 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.625130 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.625145 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:05Z","lastTransitionTime":"2025-11-24T17:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.727853 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.727927 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.727944 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.727968 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.727985 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:05Z","lastTransitionTime":"2025-11-24T17:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.830414 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.830455 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.830464 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.830489 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.830500 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:05Z","lastTransitionTime":"2025-11-24T17:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.841269 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovnkube-controller/3.log" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.845898 4808 scope.go:117] "RemoveContainer" containerID="3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222" Nov 24 17:28:05 crc kubenswrapper[4808]: E1124 17:28:05.846173 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.860815 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.873904 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.889258 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.909256 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:28:04Z\\\",\\\"message\\\":\\\"LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/catalog-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.204\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 17:28:04.116754 6882 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:28:04.116754 6882 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:28:04.116760 6882 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-ht87w in node crc\\\\nI1124 17:28:04.116764 6882 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:28:04.116770 6882 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-ht87w after 0 failed atte\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:28:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.924436 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"668040c0-c989-459b-9923-2f5a39a83958\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3462fe4e148a5c80763c265a819dcb04af31902463e9f1e34afefce17b13bbb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23297e0c2ec1bd342e97252d23b64dcf6039bcdb46579119053d8c654f953a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd30bb5725168f10a8a593c13456a101218f0bc17c492b7100470dee10fcf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.933701 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.934008 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.934098 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.934170 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.934230 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:05Z","lastTransitionTime":"2025-11-24T17:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.939628 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.953561 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.973873 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.988688 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:05 crc kubenswrapper[4808]: I1124 17:28:05.999336 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.012797 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac714762891a8496f5f9fb3d368c22acffd162a1a646871d2be0386c78b0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:56Z\\\",\\\"message\\\":\\\"2025-11-24T17:27:11+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b\\\\n2025-11-24T17:27:11+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b to /host/opt/cni/bin/\\\\n2025-11-24T17:27:11Z [verbose] multus-daemon started\\\\n2025-11-24T17:27:11Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:27:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:06Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.016324 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.016381 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.016394 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.016421 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.016436 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:06Z","lastTransitionTime":"2025-11-24T17:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.030938 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:06Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:06 crc kubenswrapper[4808]: E1124 17:28:06.031857 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:06Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.035985 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.036036 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.036049 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.036071 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.036083 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:06Z","lastTransitionTime":"2025-11-24T17:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.044558 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:06Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:06 crc kubenswrapper[4808]: E1124 17:28:06.047516 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:06Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.052076 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.052101 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.052110 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.052127 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.052138 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:06Z","lastTransitionTime":"2025-11-24T17:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.061150 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:06Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:06 crc kubenswrapper[4808]: E1124 17:28:06.068705 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:06Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.073954 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.074001 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.074015 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.074051 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.074062 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:06Z","lastTransitionTime":"2025-11-24T17:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.078855 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:06Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:06 crc kubenswrapper[4808]: E1124 17:28:06.086799 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:06Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.090531 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:06Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.090901 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.090927 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.090942 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.090961 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.090972 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:06Z","lastTransitionTime":"2025-11-24T17:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.102667 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:06Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:06 crc kubenswrapper[4808]: E1124 17:28:06.104873 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:06Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:06 crc kubenswrapper[4808]: E1124 17:28:06.105000 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.106699 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.106730 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.106744 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.106763 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.106776 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:06Z","lastTransitionTime":"2025-11-24T17:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.114164 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:06Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.209575 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.209616 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.209626 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.209641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.209651 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:06Z","lastTransitionTime":"2025-11-24T17:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.312712 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.312753 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.312763 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.312780 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.312793 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:06Z","lastTransitionTime":"2025-11-24T17:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.347144 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.347211 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.347275 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:06 crc kubenswrapper[4808]: E1124 17:28:06.347419 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.347447 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:06 crc kubenswrapper[4808]: E1124 17:28:06.347518 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:06 crc kubenswrapper[4808]: E1124 17:28:06.347582 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:06 crc kubenswrapper[4808]: E1124 17:28:06.347632 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.415226 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.415284 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.415296 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.415316 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.415330 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:06Z","lastTransitionTime":"2025-11-24T17:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.517804 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.517856 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.517869 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.517889 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.517903 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:06Z","lastTransitionTime":"2025-11-24T17:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.621651 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.621708 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.621727 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.621751 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.621764 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:06Z","lastTransitionTime":"2025-11-24T17:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.724644 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.724701 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.724712 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.724733 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.724762 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:06Z","lastTransitionTime":"2025-11-24T17:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.828273 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.828331 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.828339 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.828360 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.828374 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:06Z","lastTransitionTime":"2025-11-24T17:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.931181 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.931274 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.931287 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.931312 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:06 crc kubenswrapper[4808]: I1124 17:28:06.931326 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:06Z","lastTransitionTime":"2025-11-24T17:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.033784 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.033885 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.033898 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.033921 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.033989 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:07Z","lastTransitionTime":"2025-11-24T17:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.136614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.137375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.137393 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.137412 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.137421 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:07Z","lastTransitionTime":"2025-11-24T17:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.240830 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.240891 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.240902 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.240917 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.240928 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:07Z","lastTransitionTime":"2025-11-24T17:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.344421 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.344475 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.344486 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.344508 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.344522 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:07Z","lastTransitionTime":"2025-11-24T17:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.446798 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.446834 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.446843 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.446879 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.446888 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:07Z","lastTransitionTime":"2025-11-24T17:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.550397 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.550448 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.550459 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.550480 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.550495 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:07Z","lastTransitionTime":"2025-11-24T17:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.652925 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.652998 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.653035 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.653053 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.653066 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:07Z","lastTransitionTime":"2025-11-24T17:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.756747 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.756816 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.756833 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.756862 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.756879 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:07Z","lastTransitionTime":"2025-11-24T17:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.860444 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.860486 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.860502 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.860521 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.860534 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:07Z","lastTransitionTime":"2025-11-24T17:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.964291 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.964344 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.964362 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.964389 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:07 crc kubenswrapper[4808]: I1124 17:28:07.964407 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:07Z","lastTransitionTime":"2025-11-24T17:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.068388 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.068446 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.068462 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.068488 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.068506 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:08Z","lastTransitionTime":"2025-11-24T17:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.171802 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.171838 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.171847 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.171863 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.171873 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:08Z","lastTransitionTime":"2025-11-24T17:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.274258 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.274287 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.274295 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.274307 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.274316 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:08Z","lastTransitionTime":"2025-11-24T17:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.349359 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:08 crc kubenswrapper[4808]: E1124 17:28:08.349469 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.349625 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:08 crc kubenswrapper[4808]: E1124 17:28:08.349674 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.349784 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:08 crc kubenswrapper[4808]: E1124 17:28:08.349832 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.349961 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:08 crc kubenswrapper[4808]: E1124 17:28:08.350013 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.361531 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:08Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.373977 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:08Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.377258 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.377292 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.377303 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.377319 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.377331 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:08Z","lastTransitionTime":"2025-11-24T17:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.385388 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:08Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.401700 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:28:04Z\\\",\\\"message\\\":\\\"LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/catalog-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.204\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 17:28:04.116754 6882 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:28:04.116754 6882 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:28:04.116760 6882 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-ht87w in node crc\\\\nI1124 17:28:04.116764 6882 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:28:04.116770 6882 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-ht87w after 0 failed atte\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:28:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:08Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.413632 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"668040c0-c989-459b-9923-2f5a39a83958\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3462fe4e148a5c80763c265a819dcb04af31902463e9f1e34afefce17b13bbb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23297e0c2ec1bd342e97252d23b64dcf6039bcdb46579119053d8c654f953a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd30bb5725168f10a8a593c13456a101218f0bc17c492b7100470dee10fcf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:08Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.427096 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:08Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.437839 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:08Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.449136 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:08Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.467364 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:08Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.478733 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:08Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.481010 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.481081 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.481092 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.481113 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:08 crc kubenswrapper[4808]: I1124 17:28:08.481132 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:08Z","lastTransitionTime":"2025-11-24T17:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.145963 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:08Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.147100 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.147139 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.147150 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.147167 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.147178 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:09Z","lastTransitionTime":"2025-11-24T17:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.162772 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac714762891a8496f5f9fb3d368c22acffd162a1a646871d2be0386c78b0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:56Z\\\",\\\"message\\\":\\\"2025-11-24T17:27:11+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b\\\\n2025-11-24T17:27:11+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b to /host/opt/cni/bin/\\\\n2025-11-24T17:27:11Z [verbose] multus-daemon started\\\\n2025-11-24T17:27:11Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:27:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.179791 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.197198 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.212401 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.224562 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.250408 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.250489 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.250523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.250487 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.250548 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.250686 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:09Z","lastTransitionTime":"2025-11-24T17:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.263677 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:09Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.356194 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.356238 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.356249 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.356267 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.356279 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:09Z","lastTransitionTime":"2025-11-24T17:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.358480 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.458851 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.458893 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.458902 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.458918 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.458928 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:09Z","lastTransitionTime":"2025-11-24T17:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.560494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.560535 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.560546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.560565 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.560574 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:09Z","lastTransitionTime":"2025-11-24T17:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.662186 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.662230 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.662239 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.662254 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.662264 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:09Z","lastTransitionTime":"2025-11-24T17:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.764121 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.764184 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.764194 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.764210 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.764220 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:09Z","lastTransitionTime":"2025-11-24T17:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.866353 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.866660 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.866819 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.866929 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.867031 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:09Z","lastTransitionTime":"2025-11-24T17:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.969723 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.969766 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.969775 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.969789 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:09 crc kubenswrapper[4808]: I1124 17:28:09.969800 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:09Z","lastTransitionTime":"2025-11-24T17:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.071998 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.072759 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.072798 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.072822 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.072831 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:10Z","lastTransitionTime":"2025-11-24T17:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.175667 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.175713 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.175722 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.175737 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.175746 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:10Z","lastTransitionTime":"2025-11-24T17:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.278849 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.278899 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.278911 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.278943 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.278958 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:10Z","lastTransitionTime":"2025-11-24T17:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.346540 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.346594 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.346553 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.346699 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:10 crc kubenswrapper[4808]: E1124 17:28:10.347061 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:10 crc kubenswrapper[4808]: E1124 17:28:10.347090 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:10 crc kubenswrapper[4808]: E1124 17:28:10.347166 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:10 crc kubenswrapper[4808]: E1124 17:28:10.347347 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.381601 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.381667 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.381691 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.381718 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.381735 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:10Z","lastTransitionTime":"2025-11-24T17:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.485386 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.485447 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.485465 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.485491 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.485503 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:10Z","lastTransitionTime":"2025-11-24T17:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.595746 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.596180 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.596241 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.596270 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.596611 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:10Z","lastTransitionTime":"2025-11-24T17:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.699974 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.700050 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.700062 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.700080 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.700093 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:10Z","lastTransitionTime":"2025-11-24T17:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.803779 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.803827 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.803836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.803854 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.803867 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:10Z","lastTransitionTime":"2025-11-24T17:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.906363 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.906431 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.906449 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.906478 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:10 crc kubenswrapper[4808]: I1124 17:28:10.906500 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:10Z","lastTransitionTime":"2025-11-24T17:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.010305 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.010405 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.010425 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.010454 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.010473 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:11Z","lastTransitionTime":"2025-11-24T17:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.114153 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.114231 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.114256 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.114285 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.114303 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:11Z","lastTransitionTime":"2025-11-24T17:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.217375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.217470 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.217488 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.217516 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.217536 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:11Z","lastTransitionTime":"2025-11-24T17:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.321535 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.321631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.321657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.321695 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.321721 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:11Z","lastTransitionTime":"2025-11-24T17:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.424109 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.424167 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.424178 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.424197 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.424209 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:11Z","lastTransitionTime":"2025-11-24T17:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.527229 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.527269 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.527279 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.527296 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.527310 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:11Z","lastTransitionTime":"2025-11-24T17:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.630804 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.630847 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.630856 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.630873 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.630884 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:11Z","lastTransitionTime":"2025-11-24T17:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.734969 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.735120 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.735199 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.735273 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.735301 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:11Z","lastTransitionTime":"2025-11-24T17:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.838863 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.838943 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.838962 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.838992 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.839044 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:11Z","lastTransitionTime":"2025-11-24T17:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.943147 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.943222 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.943244 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.943275 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:11 crc kubenswrapper[4808]: I1124 17:28:11.943292 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:11Z","lastTransitionTime":"2025-11-24T17:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.048098 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.048215 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.048235 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.048266 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.048287 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:12Z","lastTransitionTime":"2025-11-24T17:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.151033 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.151106 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.151127 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.151155 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.151174 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:12Z","lastTransitionTime":"2025-11-24T17:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.180146 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.180418 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.180376512 +0000 UTC m=+148.778044334 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.254677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.254741 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.254755 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.254777 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.254789 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:12Z","lastTransitionTime":"2025-11-24T17:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.281924 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.282000 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.282046 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.282082 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.282106 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.282179 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.282155846 +0000 UTC m=+148.879823648 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.282227 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.282268 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.282226 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.282287 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.282290 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.282303 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.282361 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.282345471 +0000 UTC m=+148.880013263 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.282376 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.282370022 +0000 UTC m=+148.880037824 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.282437 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.282619 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.282584308 +0000 UTC m=+148.880252150 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.347356 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.347355 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.347372 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.347555 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.347685 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.347812 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.348085 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:12 crc kubenswrapper[4808]: E1124 17:28:12.348109 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.357932 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.357977 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.357991 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.358027 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.358045 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:12Z","lastTransitionTime":"2025-11-24T17:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.461112 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.461184 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.461205 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.461230 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.461250 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:12Z","lastTransitionTime":"2025-11-24T17:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.564625 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.564679 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.564695 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.564718 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.564733 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:12Z","lastTransitionTime":"2025-11-24T17:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.668300 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.668351 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.668361 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.668377 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.668388 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:12Z","lastTransitionTime":"2025-11-24T17:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.771115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.771169 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.771178 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.771198 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.771214 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:12Z","lastTransitionTime":"2025-11-24T17:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.875357 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.875420 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.875437 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.875458 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.875470 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:12Z","lastTransitionTime":"2025-11-24T17:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.982191 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.982700 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.982729 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.982748 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:12 crc kubenswrapper[4808]: I1124 17:28:12.982765 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:12Z","lastTransitionTime":"2025-11-24T17:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.086113 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.086172 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.086185 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.086203 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.086212 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:13Z","lastTransitionTime":"2025-11-24T17:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.188838 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.188907 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.188924 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.188949 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.188968 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:13Z","lastTransitionTime":"2025-11-24T17:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.292086 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.292158 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.292172 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.292194 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.292210 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:13Z","lastTransitionTime":"2025-11-24T17:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.395199 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.395296 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.395309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.395331 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.395345 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:13Z","lastTransitionTime":"2025-11-24T17:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.498552 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.498600 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.498609 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.498628 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.498640 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:13Z","lastTransitionTime":"2025-11-24T17:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.601540 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.601588 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.601597 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.601614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.601626 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:13Z","lastTransitionTime":"2025-11-24T17:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.705519 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.705572 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.705580 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.705601 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.705612 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:13Z","lastTransitionTime":"2025-11-24T17:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.808656 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.808699 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.808708 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.808726 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.808743 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:13Z","lastTransitionTime":"2025-11-24T17:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.912009 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.912131 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.912145 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.912170 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:13 crc kubenswrapper[4808]: I1124 17:28:13.912185 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:13Z","lastTransitionTime":"2025-11-24T17:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.015419 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.015524 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.015557 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.015594 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.015619 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:14Z","lastTransitionTime":"2025-11-24T17:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.120224 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.120302 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.120320 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.120350 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.120368 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:14Z","lastTransitionTime":"2025-11-24T17:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.223392 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.223463 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.223483 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.223508 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.223529 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:14Z","lastTransitionTime":"2025-11-24T17:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.326604 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.326663 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.326676 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.326696 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.326713 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:14Z","lastTransitionTime":"2025-11-24T17:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.347842 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:14 crc kubenswrapper[4808]: E1124 17:28:14.348056 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.348376 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:14 crc kubenswrapper[4808]: E1124 17:28:14.348505 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.348811 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.348819 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:14 crc kubenswrapper[4808]: E1124 17:28:14.349074 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:14 crc kubenswrapper[4808]: E1124 17:28:14.349204 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.431110 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.431186 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.431213 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.431242 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.431261 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:14Z","lastTransitionTime":"2025-11-24T17:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.535647 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.535725 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.535743 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.535771 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.535794 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:14Z","lastTransitionTime":"2025-11-24T17:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.639430 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.639511 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.639536 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.639569 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.639593 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:14Z","lastTransitionTime":"2025-11-24T17:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.743391 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.743460 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.743484 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.743518 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.743540 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:14Z","lastTransitionTime":"2025-11-24T17:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.848716 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.848810 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.848828 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.848856 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.848882 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:14Z","lastTransitionTime":"2025-11-24T17:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.951742 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.951793 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.951803 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.951821 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:14 crc kubenswrapper[4808]: I1124 17:28:14.951833 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:14Z","lastTransitionTime":"2025-11-24T17:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.055754 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.055809 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.055820 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.055850 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.055872 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:15Z","lastTransitionTime":"2025-11-24T17:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.158758 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.158805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.158817 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.158837 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.158850 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:15Z","lastTransitionTime":"2025-11-24T17:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.262479 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.262544 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.262557 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.262590 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.262604 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:15Z","lastTransitionTime":"2025-11-24T17:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.366115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.366289 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.366326 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.366366 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.366395 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:15Z","lastTransitionTime":"2025-11-24T17:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.470985 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.471109 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.471127 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.471160 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.471182 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:15Z","lastTransitionTime":"2025-11-24T17:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.574311 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.574383 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.574397 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.574420 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.574437 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:15Z","lastTransitionTime":"2025-11-24T17:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.677709 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.677784 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.677807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.677836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.677857 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:15Z","lastTransitionTime":"2025-11-24T17:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.782309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.782433 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.782465 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.782506 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.782551 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:15Z","lastTransitionTime":"2025-11-24T17:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.886123 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.886232 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.886245 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.886261 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.886271 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:15Z","lastTransitionTime":"2025-11-24T17:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.990594 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.990690 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.990716 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.990750 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:15 crc kubenswrapper[4808]: I1124 17:28:15.990772 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:15Z","lastTransitionTime":"2025-11-24T17:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.094287 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.094341 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.094355 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.094373 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.094385 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:16Z","lastTransitionTime":"2025-11-24T17:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.198499 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.198586 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.198605 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.198630 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.198649 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:16Z","lastTransitionTime":"2025-11-24T17:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.301275 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.301321 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.301336 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.301355 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.301369 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:16Z","lastTransitionTime":"2025-11-24T17:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.346623 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:16 crc kubenswrapper[4808]: E1124 17:28:16.346921 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.347051 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.347095 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:16 crc kubenswrapper[4808]: E1124 17:28:16.347163 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.347190 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:16 crc kubenswrapper[4808]: E1124 17:28:16.347267 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:16 crc kubenswrapper[4808]: E1124 17:28:16.347337 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.405128 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.405204 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.405225 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.405251 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.405270 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:16Z","lastTransitionTime":"2025-11-24T17:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.506823 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.506893 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.506911 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.506936 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.506953 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:16Z","lastTransitionTime":"2025-11-24T17:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:16 crc kubenswrapper[4808]: E1124 17:28:16.523209 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.527812 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.527845 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.527858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.527875 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.527888 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:16Z","lastTransitionTime":"2025-11-24T17:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:16 crc kubenswrapper[4808]: E1124 17:28:16.541832 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.545951 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.545993 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.546027 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.546045 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.546055 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:16Z","lastTransitionTime":"2025-11-24T17:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:16 crc kubenswrapper[4808]: E1124 17:28:16.560244 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.564741 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.564784 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.564799 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.564817 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.564829 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:16Z","lastTransitionTime":"2025-11-24T17:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:16 crc kubenswrapper[4808]: E1124 17:28:16.580604 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.584450 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.584562 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.584583 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.584607 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.584624 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:16Z","lastTransitionTime":"2025-11-24T17:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:16 crc kubenswrapper[4808]: E1124 17:28:16.597662 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:28:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8a5edfdd-9e82-47ae-ad4b-d584fa1b687e\\\",\\\"systemUUID\\\":\\\"c3e5358f-bf99-4bb9-b876-ca5a1061a1cf\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:16 crc kubenswrapper[4808]: E1124 17:28:16.597832 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.599873 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.599902 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.599914 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.599929 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.599940 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:16Z","lastTransitionTime":"2025-11-24T17:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.703725 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.703772 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.703790 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.703812 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.703829 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:16Z","lastTransitionTime":"2025-11-24T17:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.806521 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.806556 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.806567 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.806582 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.806593 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:16Z","lastTransitionTime":"2025-11-24T17:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.909596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.909637 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.909649 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.909667 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:16 crc kubenswrapper[4808]: I1124 17:28:16.909679 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:16Z","lastTransitionTime":"2025-11-24T17:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.012636 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.012707 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.012724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.012747 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.012762 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:17Z","lastTransitionTime":"2025-11-24T17:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.114766 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.114820 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.114836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.114854 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.114867 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:17Z","lastTransitionTime":"2025-11-24T17:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.217222 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.217264 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.217278 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.217292 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.217302 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:17Z","lastTransitionTime":"2025-11-24T17:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.320368 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.320407 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.320416 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.320431 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.320442 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:17Z","lastTransitionTime":"2025-11-24T17:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.347469 4808 scope.go:117] "RemoveContainer" containerID="3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222" Nov 24 17:28:17 crc kubenswrapper[4808]: E1124 17:28:17.347652 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.422431 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.422460 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.422469 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.422483 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.422491 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:17Z","lastTransitionTime":"2025-11-24T17:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.525154 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.525225 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.525248 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.525276 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.525296 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:17Z","lastTransitionTime":"2025-11-24T17:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.628256 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.628320 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.628334 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.628356 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.628369 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:17Z","lastTransitionTime":"2025-11-24T17:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.731470 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.731526 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.731537 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.731555 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.731566 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:17Z","lastTransitionTime":"2025-11-24T17:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.834651 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.834698 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.834707 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.834723 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.834734 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:17Z","lastTransitionTime":"2025-11-24T17:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.938091 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.938142 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.938150 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.938168 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:17 crc kubenswrapper[4808]: I1124 17:28:17.938179 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:17Z","lastTransitionTime":"2025-11-24T17:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.040496 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.040553 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.040565 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.040585 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.040602 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:18Z","lastTransitionTime":"2025-11-24T17:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.143309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.143416 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.143427 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.143444 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.143454 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:18Z","lastTransitionTime":"2025-11-24T17:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.248321 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.248428 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.248452 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.248485 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.248507 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:18Z","lastTransitionTime":"2025-11-24T17:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.347406 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.347421 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.347488 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:18 crc kubenswrapper[4808]: E1124 17:28:18.347742 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:18 crc kubenswrapper[4808]: E1124 17:28:18.348102 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.348308 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:18 crc kubenswrapper[4808]: E1124 17:28:18.348442 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:18 crc kubenswrapper[4808]: E1124 17:28:18.348508 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.351989 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.352104 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.352125 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.352154 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.352175 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:18Z","lastTransitionTime":"2025-11-24T17:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.369879 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.384913 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7cbtg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ded1d94f-dc75-45de-b921-a66747cd3f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f49987ada4d963ce5f8d8b97411e873494f2d11fa82d29e7451fba963c0e2b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p7wjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7cbtg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.402143 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q5m49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a192cd6-853d-4643-855d-ac61168b1cda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac714762891a8496f5f9fb3d368c22acffd162a1a646871d2be0386c78b0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:27:56Z\\\",\\\"message\\\":\\\"2025-11-24T17:27:11+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b\\\\n2025-11-24T17:27:11+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3d851e4f-5f7c-42ff-9e88-093d4170920b to /host/opt/cni/bin/\\\\n2025-11-24T17:27:11Z [verbose] multus-daemon started\\\\n2025-11-24T17:27:11Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:27:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6tl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q5m49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.421036 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe424e29-3d8f-49e3-ac0c-33dc63f93cf8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c97cbd67a2ec3bb20e443b84b13299b684f2cd1dcb58f7a4d88a39851e5db69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b1386713b1402ac1ec434c1fd35f51de566eb751bf4ae9a1a387484c14711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b03102d5b296885bf8829e6707393cd1f4fe9e70d4452317a5b61c60391acbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be67ed36a405ea0abcebd8c6098f6edc5e78c876ff34523f4e946639d982bacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f27f15f2f5aa311f56f0c5378a6191b049623df4c1e64b4b2ca1b1ca9ebb50e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83c520877df0c9f922149280b2dafb33a88bdd1daa09dfd1894a12d803326c93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f23afabd8fe38486905f266f955136416d9ea10a4755f6726aa9be0920639a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zgw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.435622 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c26458eb-7f6c-42cb-b6c5-90de6756a994\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8v2q4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.458784 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8330036-fa01-4ca9-90e3-62feb6c78fb0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b773d5f435f5667a1fb8f1faa77a467eb5a805cf9290c8c37cc5a50b81574214\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb1d3948777c99fe49eee2cf342bfdb814c54d6dae2d171b91fe9e7f846875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81b9c8c41df6aba5ce4aeef7ef471d6b7a7ebbd292d6abe6635c3a3fb2f438\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://872f3cd55a0aed5a0dcc229792b3efea67dfe066e0f54ab783aec30643cbf8f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.461391 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.461433 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.461442 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.461479 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.461491 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:18Z","lastTransitionTime":"2025-11-24T17:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.492531 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"264b09f4-bd1f-4eda-8ad3-e39c22ee317c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de80d3ff4d089ffa7a1d8be361ee310f663c2cf55f527f3218abc63315849f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7fcf4a9682065f0729059e1ee517cc544e8fbad243513240b592888780bbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e980686ee58ef4c7448b4cac9d9101a3018bacf851e4361789ff656e255da6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://236a6dadeffdb2723bd8e448c9b7322f7db4d7d1bbdb5652dd0c25a7c2589c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5a85f41881dec1573269db856c8747e3ccde1197f39c6164dd563df941e8c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0812f88dca31058f163489c58ecb43145ba895cf51f0d3fe1a19667b785859\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae9d9888457e536258a0de0e5a9df5a847eb7115f1cfc63fd67d77f18f829816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3082b9d80b046167e34114d06a3f77c69aff848c6550c461cab290356c7683e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.508828 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k9pwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c07994f6-627f-4641-b164-ae81aac54782\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d900f7d2dc884524ca2e33c8b27494c30f6f3289129ea53aa60a98922f4850ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qr4gx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k9pwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.523623 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bde5653fc43fc515535327ab1620ef73c9acb22769456190414d49616a10c3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhss2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ht87w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.539714 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce6613f-d181-4eb4-9cfc-4762964eb8d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc9eed56e27beeb4559db55927bf9d9cd8bf6a3b7cbbf3f4d2c24437fe76e24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da811c4233c48ed57e9514f3470a0ac783242b8a251ee5b0b9e4dc75c2bcecfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k22s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p5b8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.557270 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ea6e4a3be04e50fc22f21e3629fee47c9395c9090a3e4b6c1d0c049eef1c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e60de74ebf1f1a66ab394c6bf685a24769e6e5e0a0283c064121b12cf4f1723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.563766 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.563794 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.563803 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.563816 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.563825 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:18Z","lastTransitionTime":"2025-11-24T17:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.571805 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.586102 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.599421 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688cff727594968afc08c37e24778d0bc8ed20f46e494a29f1973c279d2b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.621043 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4caad6d6-912a-4441-aa90-38a0e29591f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:28:04Z\\\",\\\"message\\\":\\\"LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/catalog-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.204\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 17:28:04.116754 6882 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:28:04.116754 6882 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 17:28:04.116760 6882 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-ht87w in node crc\\\\nI1124 17:28:04.116764 6882 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:28:04.116770 6882 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-ht87w after 0 failed atte\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:28:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrc9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:27:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5b2z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.635344 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d61802a8-4bb2-44f8-9546-ff56e981a52b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cfcd462f857faabd6f524d8ac5037283e272ba0705749a8df1a058a1279e753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1484dcf3f52f8d8388542b94e030ddab6f4fa2921d97269650d616ac488bca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1484dcf3f52f8d8388542b94e030ddab6f4fa2921d97269650d616ac488bca3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.650012 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0868d7c-6a52-46e9-8f59-b99565eff4ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9c9a9fcf63171e8b4faa204e030f90eabcde888caafda228005211e96f8c9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23947bdbb420f073fc7201991d502f58927effe3d253b2a0c4f7aeaa9a4116a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19198f5752996897a98e789ed46c00e3236a2bd061018380d3e6050875f79538\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcc14b3fc5dd2d11cf6ea56d6d5283f9bc07fa46c667d3f76c19b8d83ee60df9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2de685edf03135bc55d74163c01e7ccbfd01d9259d6faea168ff87fbaafceb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:27:08Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:27:08.539352 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1124 17:27:08.539348 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:27:08.539356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:27:08.539374 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 17:27:08.539379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:27:08.539382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 17:27:08.541765 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541795 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 17:27:08.541817 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1124 17:27:08.541831 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 17:27:08.541860 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 17:27:08.541927 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 17:27:08.541942 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1124 17:27:08.541967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:27:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a85e78932c5ae6e61f91d31b35e42f296adb3452676626e960613eaaaf4689b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a09ec3a407d1cc2d4733fdf426869a3616b194265fa871b26f0b7fbff4b88ce6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.663221 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"668040c0-c989-459b-9923-2f5a39a83958\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3462fe4e148a5c80763c265a819dcb04af31902463e9f1e34afefce17b13bbb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23297e0c2ec1bd342e97252d23b64dcf6039bcdb46579119053d8c654f953a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd30bb5725168f10a8a593c13456a101218f0bc17c492b7100470dee10fcf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:26:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://015191b839591ac10dda466537c1c3b0f6257984187936f1d9df272493bd2c11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:26:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:26:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:26:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.667214 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.667267 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.667280 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.667301 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.667315 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:18Z","lastTransitionTime":"2025-11-24T17:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.680403 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:27:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7c3d73a2be1d8a15f97849b1074e0c65ace564b915563eda76636b094669ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:27:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:28:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.776870 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.776941 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.776958 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.776991 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.777009 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:18Z","lastTransitionTime":"2025-11-24T17:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.879472 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.879524 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.879536 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.879558 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.879575 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:18Z","lastTransitionTime":"2025-11-24T17:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.982527 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.982587 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.982642 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.982725 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:18 crc kubenswrapper[4808]: I1124 17:28:18.982746 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:18Z","lastTransitionTime":"2025-11-24T17:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.085978 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.086107 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.086134 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.086170 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.086195 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:19Z","lastTransitionTime":"2025-11-24T17:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.189387 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.189451 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.189467 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.189492 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.189509 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:19Z","lastTransitionTime":"2025-11-24T17:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.292947 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.293056 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.293076 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.293110 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.293135 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:19Z","lastTransitionTime":"2025-11-24T17:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.396032 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.396081 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.396092 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.396107 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.396118 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:19Z","lastTransitionTime":"2025-11-24T17:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.499084 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.499164 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.499175 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.499196 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.499206 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:19Z","lastTransitionTime":"2025-11-24T17:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.602169 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.602226 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.602236 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.602252 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.602265 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:19Z","lastTransitionTime":"2025-11-24T17:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.704860 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.704934 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.704950 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.704973 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.704990 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:19Z","lastTransitionTime":"2025-11-24T17:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.807427 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.807505 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.807514 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.807529 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.807540 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:19Z","lastTransitionTime":"2025-11-24T17:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.910799 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.910845 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.910854 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.910867 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:19 crc kubenswrapper[4808]: I1124 17:28:19.910878 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:19Z","lastTransitionTime":"2025-11-24T17:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.013483 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.013528 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.013538 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.013553 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.013562 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:20Z","lastTransitionTime":"2025-11-24T17:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.116823 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.116890 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.116909 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.116935 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.116948 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:20Z","lastTransitionTime":"2025-11-24T17:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.220176 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.220237 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.220250 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.220274 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.220291 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:20Z","lastTransitionTime":"2025-11-24T17:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.323060 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.323097 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.323107 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.323123 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.323137 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:20Z","lastTransitionTime":"2025-11-24T17:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.346773 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.346772 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.346784 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:20 crc kubenswrapper[4808]: E1124 17:28:20.347235 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:20 crc kubenswrapper[4808]: E1124 17:28:20.346982 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:20 crc kubenswrapper[4808]: E1124 17:28:20.347259 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.346796 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:20 crc kubenswrapper[4808]: E1124 17:28:20.347414 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.427196 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.427250 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.427263 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.427286 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.427303 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:20Z","lastTransitionTime":"2025-11-24T17:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.531312 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.531368 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.531381 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.531401 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.531417 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:20Z","lastTransitionTime":"2025-11-24T17:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.634843 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.634909 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.634926 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.634955 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.634973 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:20Z","lastTransitionTime":"2025-11-24T17:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.737741 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.737788 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.737797 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.737815 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.737825 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:20Z","lastTransitionTime":"2025-11-24T17:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.840744 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.840792 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.840802 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.840818 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.840829 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:20Z","lastTransitionTime":"2025-11-24T17:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.946872 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.946928 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.946937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.946954 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:20 crc kubenswrapper[4808]: I1124 17:28:20.946964 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:20Z","lastTransitionTime":"2025-11-24T17:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.050453 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.050559 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.050573 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.050593 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.050606 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:21Z","lastTransitionTime":"2025-11-24T17:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.153628 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.153681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.153694 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.153719 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.153732 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:21Z","lastTransitionTime":"2025-11-24T17:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.256336 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.256691 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.256843 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.256914 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.257265 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:21Z","lastTransitionTime":"2025-11-24T17:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.360143 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.360198 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.360213 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.360235 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.360250 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:21Z","lastTransitionTime":"2025-11-24T17:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.463331 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.463913 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.464133 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.464299 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.464443 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:21Z","lastTransitionTime":"2025-11-24T17:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.567971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.568056 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.568072 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.568096 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.568115 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:21Z","lastTransitionTime":"2025-11-24T17:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.671605 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.671664 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.671680 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.671709 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.671729 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:21Z","lastTransitionTime":"2025-11-24T17:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.774739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.774784 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.774796 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.774811 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.774822 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:21Z","lastTransitionTime":"2025-11-24T17:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.877611 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.877649 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.877660 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.877677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.877688 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:21Z","lastTransitionTime":"2025-11-24T17:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.980534 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.980575 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.980584 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.980597 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:21 crc kubenswrapper[4808]: I1124 17:28:21.980606 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:21Z","lastTransitionTime":"2025-11-24T17:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.082986 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.083120 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.083135 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.083162 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.083179 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:22Z","lastTransitionTime":"2025-11-24T17:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.185774 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.185852 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.185868 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.185895 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.185916 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:22Z","lastTransitionTime":"2025-11-24T17:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.288826 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.288875 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.288885 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.288908 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.288918 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:22Z","lastTransitionTime":"2025-11-24T17:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.346819 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.346873 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.346916 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.347104 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:22 crc kubenswrapper[4808]: E1124 17:28:22.347233 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:22 crc kubenswrapper[4808]: E1124 17:28:22.347353 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:22 crc kubenswrapper[4808]: E1124 17:28:22.347496 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:22 crc kubenswrapper[4808]: E1124 17:28:22.347617 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.392075 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.392134 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.392150 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.392175 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.392191 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:22Z","lastTransitionTime":"2025-11-24T17:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.496294 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.496368 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.496381 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.496402 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.496444 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:22Z","lastTransitionTime":"2025-11-24T17:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.600830 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.600920 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.600945 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.600978 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.600997 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:22Z","lastTransitionTime":"2025-11-24T17:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.704175 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.704218 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.704227 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.704250 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.704261 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:22Z","lastTransitionTime":"2025-11-24T17:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.807514 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.807656 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.807678 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.807707 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.807727 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:22Z","lastTransitionTime":"2025-11-24T17:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.911164 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.911229 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.911247 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.911272 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:22 crc kubenswrapper[4808]: I1124 17:28:22.911289 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:22Z","lastTransitionTime":"2025-11-24T17:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.013708 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.013756 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.013773 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.013795 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.013813 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:23Z","lastTransitionTime":"2025-11-24T17:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.117396 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.117466 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.117482 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.117510 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.117547 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:23Z","lastTransitionTime":"2025-11-24T17:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.220048 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.220112 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.220124 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.220146 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.220161 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:23Z","lastTransitionTime":"2025-11-24T17:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.323798 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.323859 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.323874 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.323893 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.323905 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:23Z","lastTransitionTime":"2025-11-24T17:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.426498 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.426565 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.426581 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.426605 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.426622 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:23Z","lastTransitionTime":"2025-11-24T17:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.530521 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.530598 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.530622 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.530654 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.530677 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:23Z","lastTransitionTime":"2025-11-24T17:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.633530 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.633573 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.633581 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.633596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.633605 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:23Z","lastTransitionTime":"2025-11-24T17:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.736787 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.736867 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.736893 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.736929 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.736953 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:23Z","lastTransitionTime":"2025-11-24T17:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.839929 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.839992 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.840009 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.840059 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.840077 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:23Z","lastTransitionTime":"2025-11-24T17:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.944310 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.944379 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.944396 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.944420 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:23 crc kubenswrapper[4808]: I1124 17:28:23.944435 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:23Z","lastTransitionTime":"2025-11-24T17:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.048132 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.048209 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.048227 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.048257 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.048281 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:24Z","lastTransitionTime":"2025-11-24T17:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.150843 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.150903 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.150920 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.150951 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.150969 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:24Z","lastTransitionTime":"2025-11-24T17:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.255485 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.255542 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.255560 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.255586 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.255608 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:24Z","lastTransitionTime":"2025-11-24T17:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.347270 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.347269 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.347370 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.347393 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:24 crc kubenswrapper[4808]: E1124 17:28:24.347626 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:24 crc kubenswrapper[4808]: E1124 17:28:24.347801 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:24 crc kubenswrapper[4808]: E1124 17:28:24.348072 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:24 crc kubenswrapper[4808]: E1124 17:28:24.348246 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.358633 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.358717 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.358740 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.358771 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.358800 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:24Z","lastTransitionTime":"2025-11-24T17:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.461628 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.461676 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.461688 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.461707 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.461720 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:24Z","lastTransitionTime":"2025-11-24T17:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.564518 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.564564 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.564576 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.564595 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.564607 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:24Z","lastTransitionTime":"2025-11-24T17:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.667608 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.667660 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.667671 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.667689 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.667701 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:24Z","lastTransitionTime":"2025-11-24T17:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.771338 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.771413 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.771433 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.771841 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.771867 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:24Z","lastTransitionTime":"2025-11-24T17:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.874872 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.875161 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.875269 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.875373 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.875615 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:24Z","lastTransitionTime":"2025-11-24T17:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.979523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.979591 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.979609 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.979635 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:24 crc kubenswrapper[4808]: I1124 17:28:24.979655 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:24Z","lastTransitionTime":"2025-11-24T17:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.082778 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.082846 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.082887 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.082919 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.082943 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:25Z","lastTransitionTime":"2025-11-24T17:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.186598 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.186685 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.186707 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.186742 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.186768 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:25Z","lastTransitionTime":"2025-11-24T17:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.289316 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.289403 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.289440 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.289473 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.289497 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:25Z","lastTransitionTime":"2025-11-24T17:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.392305 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.392378 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.392399 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.392428 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.392449 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:25Z","lastTransitionTime":"2025-11-24T17:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.495839 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.495924 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.495953 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.495988 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.496063 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:25Z","lastTransitionTime":"2025-11-24T17:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.598511 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.598591 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.598601 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.598620 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.598657 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:25Z","lastTransitionTime":"2025-11-24T17:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.701879 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.701971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.701990 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.702061 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.702096 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:25Z","lastTransitionTime":"2025-11-24T17:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.806179 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.806255 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.806273 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.806304 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.806326 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:25Z","lastTransitionTime":"2025-11-24T17:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.908809 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.908882 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.908902 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.908929 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:25 crc kubenswrapper[4808]: I1124 17:28:25.908946 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:25Z","lastTransitionTime":"2025-11-24T17:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.012682 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.012778 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.012792 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.012809 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.012822 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:26Z","lastTransitionTime":"2025-11-24T17:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.116158 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.116207 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.116219 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.116238 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.116251 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:26Z","lastTransitionTime":"2025-11-24T17:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.218886 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.218993 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.219006 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.219047 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.219062 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:26Z","lastTransitionTime":"2025-11-24T17:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.322433 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.322501 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.322516 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.322536 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.322550 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:26Z","lastTransitionTime":"2025-11-24T17:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.347416 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.347428 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:26 crc kubenswrapper[4808]: E1124 17:28:26.347632 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.347534 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:26 crc kubenswrapper[4808]: E1124 17:28:26.347744 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:26 crc kubenswrapper[4808]: E1124 17:28:26.347821 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.348216 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:26 crc kubenswrapper[4808]: E1124 17:28:26.348533 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.363601 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs\") pod \"network-metrics-daemon-8v2q4\" (UID: \"c26458eb-7f6c-42cb-b6c5-90de6756a994\") " pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:26 crc kubenswrapper[4808]: E1124 17:28:26.364205 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:28:26 crc kubenswrapper[4808]: E1124 17:28:26.364443 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs podName:c26458eb-7f6c-42cb-b6c5-90de6756a994 nodeName:}" failed. No retries permitted until 2025-11-24 17:29:30.364351934 +0000 UTC m=+162.962019746 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs") pod "network-metrics-daemon-8v2q4" (UID: "c26458eb-7f6c-42cb-b6c5-90de6756a994") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.425054 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.425113 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.425124 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.425141 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.425153 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:26Z","lastTransitionTime":"2025-11-24T17:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.528319 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.528419 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.528449 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.528467 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.528482 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:26Z","lastTransitionTime":"2025-11-24T17:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.622916 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.622970 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.622988 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.623010 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.623041 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:28:26Z","lastTransitionTime":"2025-11-24T17:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.682934 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk"] Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.683581 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.686759 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.687029 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.687316 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.691717 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.736698 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-zg5rq" podStartSLOduration=78.736656906 podStartE2EDuration="1m18.736656906s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:28:26.723648901 +0000 UTC m=+99.321316713" watchObservedRunningTime="2025-11-24 17:28:26.736656906 +0000 UTC m=+99.334324748" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.757291 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=73.75725958 podStartE2EDuration="1m13.75725958s" podCreationTimestamp="2025-11-24 17:27:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:28:26.757249419 +0000 UTC m=+99.354917261" watchObservedRunningTime="2025-11-24 17:28:26.75725958 +0000 UTC m=+99.354927392" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.771383 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f02fc106-7811-4d65-8413-557bd93b66ec-service-ca\") pod \"cluster-version-operator-5c965bbfc6-r9dbk\" (UID: \"f02fc106-7811-4d65-8413-557bd93b66ec\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.771446 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/f02fc106-7811-4d65-8413-557bd93b66ec-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-r9dbk\" (UID: \"f02fc106-7811-4d65-8413-557bd93b66ec\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.771473 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f02fc106-7811-4d65-8413-557bd93b66ec-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-r9dbk\" (UID: \"f02fc106-7811-4d65-8413-557bd93b66ec\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.771514 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/f02fc106-7811-4d65-8413-557bd93b66ec-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-r9dbk\" (UID: \"f02fc106-7811-4d65-8413-557bd93b66ec\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.771560 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f02fc106-7811-4d65-8413-557bd93b66ec-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-r9dbk\" (UID: \"f02fc106-7811-4d65-8413-557bd93b66ec\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.808980 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=75.80894441 podStartE2EDuration="1m15.80894441s" podCreationTimestamp="2025-11-24 17:27:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:28:26.791842297 +0000 UTC m=+99.389510109" watchObservedRunningTime="2025-11-24 17:28:26.80894441 +0000 UTC m=+99.406612232" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.822309 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-7cbtg" podStartSLOduration=79.822287314 podStartE2EDuration="1m19.822287314s" podCreationTimestamp="2025-11-24 17:27:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:28:26.821915743 +0000 UTC m=+99.419583545" watchObservedRunningTime="2025-11-24 17:28:26.822287314 +0000 UTC m=+99.419955136" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.854168 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-q5m49" podStartSLOduration=78.85413595200001 podStartE2EDuration="1m18.854135952s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:28:26.839718947 +0000 UTC m=+99.437386769" watchObservedRunningTime="2025-11-24 17:28:26.854135952 +0000 UTC m=+99.451803754" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.872594 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f02fc106-7811-4d65-8413-557bd93b66ec-service-ca\") pod \"cluster-version-operator-5c965bbfc6-r9dbk\" (UID: \"f02fc106-7811-4d65-8413-557bd93b66ec\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.872658 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/f02fc106-7811-4d65-8413-557bd93b66ec-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-r9dbk\" (UID: \"f02fc106-7811-4d65-8413-557bd93b66ec\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.872687 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f02fc106-7811-4d65-8413-557bd93b66ec-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-r9dbk\" (UID: \"f02fc106-7811-4d65-8413-557bd93b66ec\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.872732 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/f02fc106-7811-4d65-8413-557bd93b66ec-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-r9dbk\" (UID: \"f02fc106-7811-4d65-8413-557bd93b66ec\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.872774 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f02fc106-7811-4d65-8413-557bd93b66ec-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-r9dbk\" (UID: \"f02fc106-7811-4d65-8413-557bd93b66ec\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.872791 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/f02fc106-7811-4d65-8413-557bd93b66ec-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-r9dbk\" (UID: \"f02fc106-7811-4d65-8413-557bd93b66ec\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.872827 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/f02fc106-7811-4d65-8413-557bd93b66ec-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-r9dbk\" (UID: \"f02fc106-7811-4d65-8413-557bd93b66ec\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.873617 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f02fc106-7811-4d65-8413-557bd93b66ec-service-ca\") pod \"cluster-version-operator-5c965bbfc6-r9dbk\" (UID: \"f02fc106-7811-4d65-8413-557bd93b66ec\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.879775 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f02fc106-7811-4d65-8413-557bd93b66ec-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-r9dbk\" (UID: \"f02fc106-7811-4d65-8413-557bd93b66ec\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.900907 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f02fc106-7811-4d65-8413-557bd93b66ec-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-r9dbk\" (UID: \"f02fc106-7811-4d65-8413-557bd93b66ec\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.918898 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-k9pwr" podStartSLOduration=79.918869678 podStartE2EDuration="1m19.918869678s" podCreationTimestamp="2025-11-24 17:27:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:28:26.907416128 +0000 UTC m=+99.505083930" watchObservedRunningTime="2025-11-24 17:28:26.918869678 +0000 UTC m=+99.516537480" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.919331 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podStartSLOduration=79.919324801 podStartE2EDuration="1m19.919324801s" podCreationTimestamp="2025-11-24 17:27:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:28:26.918972691 +0000 UTC m=+99.516640493" watchObservedRunningTime="2025-11-24 17:28:26.919324801 +0000 UTC m=+99.516992593" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.935278 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p5b8b" podStartSLOduration=78.93525311 podStartE2EDuration="1m18.93525311s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:28:26.934931101 +0000 UTC m=+99.532598893" watchObservedRunningTime="2025-11-24 17:28:26.93525311 +0000 UTC m=+99.532920912" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.949697 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=17.949669926 podStartE2EDuration="17.949669926s" podCreationTimestamp="2025-11-24 17:28:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:28:26.948803401 +0000 UTC m=+99.546471203" watchObservedRunningTime="2025-11-24 17:28:26.949669926 +0000 UTC m=+99.547337728" Nov 24 17:28:26 crc kubenswrapper[4808]: I1124 17:28:26.965909 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.965887843 podStartE2EDuration="1m18.965887843s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:28:26.965511933 +0000 UTC m=+99.563179745" watchObservedRunningTime="2025-11-24 17:28:26.965887843 +0000 UTC m=+99.563555645" Nov 24 17:28:27 crc kubenswrapper[4808]: I1124 17:28:27.010011 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" Nov 24 17:28:27 crc kubenswrapper[4808]: I1124 17:28:27.038954 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=47.038931639 podStartE2EDuration="47.038931639s" podCreationTimestamp="2025-11-24 17:27:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:28:27.038768934 +0000 UTC m=+99.636436736" watchObservedRunningTime="2025-11-24 17:28:27.038931639 +0000 UTC m=+99.636599431" Nov 24 17:28:27 crc kubenswrapper[4808]: I1124 17:28:27.208378 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" event={"ID":"f02fc106-7811-4d65-8413-557bd93b66ec","Type":"ContainerStarted","Data":"81b0e74b8a61457e936c86c4c1091d7cbb03b71a7270e023996ffd0b7c00ee53"} Nov 24 17:28:27 crc kubenswrapper[4808]: I1124 17:28:27.208773 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" event={"ID":"f02fc106-7811-4d65-8413-557bd93b66ec","Type":"ContainerStarted","Data":"321c54a1bb8f2e42b944e359f9d3f6f1a8eb36e51029ba7202b8798f92462ea3"} Nov 24 17:28:27 crc kubenswrapper[4808]: I1124 17:28:27.227277 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r9dbk" podStartSLOduration=79.227252197 podStartE2EDuration="1m19.227252197s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:28:27.226127345 +0000 UTC m=+99.823795147" watchObservedRunningTime="2025-11-24 17:28:27.227252197 +0000 UTC m=+99.824920009" Nov 24 17:28:28 crc kubenswrapper[4808]: I1124 17:28:28.346529 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:28 crc kubenswrapper[4808]: I1124 17:28:28.346615 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:28 crc kubenswrapper[4808]: E1124 17:28:28.347627 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:28 crc kubenswrapper[4808]: I1124 17:28:28.347646 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:28 crc kubenswrapper[4808]: I1124 17:28:28.347733 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:28 crc kubenswrapper[4808]: E1124 17:28:28.347782 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:28 crc kubenswrapper[4808]: E1124 17:28:28.347970 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:28 crc kubenswrapper[4808]: E1124 17:28:28.348105 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:30 crc kubenswrapper[4808]: I1124 17:28:30.347427 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:30 crc kubenswrapper[4808]: I1124 17:28:30.347448 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:30 crc kubenswrapper[4808]: I1124 17:28:30.347747 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:30 crc kubenswrapper[4808]: I1124 17:28:30.347805 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:30 crc kubenswrapper[4808]: E1124 17:28:30.348070 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:30 crc kubenswrapper[4808]: E1124 17:28:30.348469 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:30 crc kubenswrapper[4808]: E1124 17:28:30.348577 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:30 crc kubenswrapper[4808]: E1124 17:28:30.349453 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:30 crc kubenswrapper[4808]: I1124 17:28:30.349916 4808 scope.go:117] "RemoveContainer" containerID="3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222" Nov 24 17:28:30 crc kubenswrapper[4808]: E1124 17:28:30.350272 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" Nov 24 17:28:32 crc kubenswrapper[4808]: I1124 17:28:32.347280 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:32 crc kubenswrapper[4808]: E1124 17:28:32.347447 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:32 crc kubenswrapper[4808]: I1124 17:28:32.347549 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:32 crc kubenswrapper[4808]: I1124 17:28:32.347569 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:32 crc kubenswrapper[4808]: I1124 17:28:32.347853 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:32 crc kubenswrapper[4808]: E1124 17:28:32.347843 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:32 crc kubenswrapper[4808]: E1124 17:28:32.347957 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:32 crc kubenswrapper[4808]: E1124 17:28:32.348064 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:34 crc kubenswrapper[4808]: I1124 17:28:34.346943 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:34 crc kubenswrapper[4808]: I1124 17:28:34.347142 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:34 crc kubenswrapper[4808]: E1124 17:28:34.347177 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:34 crc kubenswrapper[4808]: I1124 17:28:34.347216 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:34 crc kubenswrapper[4808]: I1124 17:28:34.347269 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:34 crc kubenswrapper[4808]: E1124 17:28:34.347446 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:34 crc kubenswrapper[4808]: E1124 17:28:34.347640 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:34 crc kubenswrapper[4808]: E1124 17:28:34.347782 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:36 crc kubenswrapper[4808]: I1124 17:28:36.347184 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:36 crc kubenswrapper[4808]: I1124 17:28:36.347252 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:36 crc kubenswrapper[4808]: E1124 17:28:36.348171 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:36 crc kubenswrapper[4808]: I1124 17:28:36.347592 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:36 crc kubenswrapper[4808]: E1124 17:28:36.348497 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:36 crc kubenswrapper[4808]: E1124 17:28:36.348235 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:36 crc kubenswrapper[4808]: I1124 17:28:36.347274 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:36 crc kubenswrapper[4808]: E1124 17:28:36.348819 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:38 crc kubenswrapper[4808]: I1124 17:28:38.347285 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:38 crc kubenswrapper[4808]: E1124 17:28:38.347512 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:38 crc kubenswrapper[4808]: I1124 17:28:38.350161 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:38 crc kubenswrapper[4808]: E1124 17:28:38.350275 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:38 crc kubenswrapper[4808]: I1124 17:28:38.350502 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:38 crc kubenswrapper[4808]: E1124 17:28:38.350611 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:38 crc kubenswrapper[4808]: I1124 17:28:38.350922 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:38 crc kubenswrapper[4808]: E1124 17:28:38.351059 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:40 crc kubenswrapper[4808]: I1124 17:28:40.347047 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:40 crc kubenswrapper[4808]: I1124 17:28:40.347106 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:40 crc kubenswrapper[4808]: I1124 17:28:40.347163 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:40 crc kubenswrapper[4808]: E1124 17:28:40.347199 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:40 crc kubenswrapper[4808]: I1124 17:28:40.347136 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:40 crc kubenswrapper[4808]: E1124 17:28:40.347397 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:40 crc kubenswrapper[4808]: E1124 17:28:40.347697 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:40 crc kubenswrapper[4808]: E1124 17:28:40.347564 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:42 crc kubenswrapper[4808]: I1124 17:28:42.310505 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q5m49_6a192cd6-853d-4643-855d-ac61168b1cda/kube-multus/1.log" Nov 24 17:28:42 crc kubenswrapper[4808]: I1124 17:28:42.311735 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q5m49_6a192cd6-853d-4643-855d-ac61168b1cda/kube-multus/0.log" Nov 24 17:28:42 crc kubenswrapper[4808]: I1124 17:28:42.311914 4808 generic.go:334] "Generic (PLEG): container finished" podID="6a192cd6-853d-4643-855d-ac61168b1cda" containerID="3ac714762891a8496f5f9fb3d368c22acffd162a1a646871d2be0386c78b0a21" exitCode=1 Nov 24 17:28:42 crc kubenswrapper[4808]: I1124 17:28:42.312054 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q5m49" event={"ID":"6a192cd6-853d-4643-855d-ac61168b1cda","Type":"ContainerDied","Data":"3ac714762891a8496f5f9fb3d368c22acffd162a1a646871d2be0386c78b0a21"} Nov 24 17:28:42 crc kubenswrapper[4808]: I1124 17:28:42.312211 4808 scope.go:117] "RemoveContainer" containerID="e4c9e5ca9a132452d560ab9b590cf62c0e1f3f207f9f5aa09cc95900e216b147" Nov 24 17:28:42 crc kubenswrapper[4808]: I1124 17:28:42.313628 4808 scope.go:117] "RemoveContainer" containerID="3ac714762891a8496f5f9fb3d368c22acffd162a1a646871d2be0386c78b0a21" Nov 24 17:28:42 crc kubenswrapper[4808]: E1124 17:28:42.313942 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-q5m49_openshift-multus(6a192cd6-853d-4643-855d-ac61168b1cda)\"" pod="openshift-multus/multus-q5m49" podUID="6a192cd6-853d-4643-855d-ac61168b1cda" Nov 24 17:28:42 crc kubenswrapper[4808]: I1124 17:28:42.347828 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:42 crc kubenswrapper[4808]: I1124 17:28:42.347859 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:42 crc kubenswrapper[4808]: I1124 17:28:42.348047 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:42 crc kubenswrapper[4808]: E1124 17:28:42.348201 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:42 crc kubenswrapper[4808]: I1124 17:28:42.348473 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:42 crc kubenswrapper[4808]: E1124 17:28:42.349107 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:42 crc kubenswrapper[4808]: E1124 17:28:42.349327 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:42 crc kubenswrapper[4808]: E1124 17:28:42.349400 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:42 crc kubenswrapper[4808]: I1124 17:28:42.349695 4808 scope.go:117] "RemoveContainer" containerID="3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222" Nov 24 17:28:42 crc kubenswrapper[4808]: E1124 17:28:42.350106 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5b2z8_openshift-ovn-kubernetes(4caad6d6-912a-4441-aa90-38a0e29591f5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" Nov 24 17:28:43 crc kubenswrapper[4808]: I1124 17:28:43.317674 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q5m49_6a192cd6-853d-4643-855d-ac61168b1cda/kube-multus/1.log" Nov 24 17:28:43 crc kubenswrapper[4808]: I1124 17:28:43.403784 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:43 crc kubenswrapper[4808]: E1124 17:28:43.404612 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:44 crc kubenswrapper[4808]: I1124 17:28:44.346497 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:44 crc kubenswrapper[4808]: E1124 17:28:44.346688 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:44 crc kubenswrapper[4808]: I1124 17:28:44.346744 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:44 crc kubenswrapper[4808]: E1124 17:28:44.346926 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:44 crc kubenswrapper[4808]: I1124 17:28:44.347001 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:44 crc kubenswrapper[4808]: E1124 17:28:44.347126 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:45 crc kubenswrapper[4808]: I1124 17:28:45.346700 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:45 crc kubenswrapper[4808]: E1124 17:28:45.346925 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:46 crc kubenswrapper[4808]: I1124 17:28:46.347324 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:46 crc kubenswrapper[4808]: I1124 17:28:46.347353 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:46 crc kubenswrapper[4808]: I1124 17:28:46.347369 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:46 crc kubenswrapper[4808]: E1124 17:28:46.347490 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:46 crc kubenswrapper[4808]: E1124 17:28:46.347601 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:46 crc kubenswrapper[4808]: E1124 17:28:46.347732 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:47 crc kubenswrapper[4808]: I1124 17:28:47.346651 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:47 crc kubenswrapper[4808]: E1124 17:28:47.347333 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:48 crc kubenswrapper[4808]: I1124 17:28:48.347275 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:48 crc kubenswrapper[4808]: E1124 17:28:48.348946 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:48 crc kubenswrapper[4808]: I1124 17:28:48.349164 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:48 crc kubenswrapper[4808]: I1124 17:28:48.349297 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:48 crc kubenswrapper[4808]: E1124 17:28:48.349413 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:48 crc kubenswrapper[4808]: E1124 17:28:48.349751 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:48 crc kubenswrapper[4808]: E1124 17:28:48.370813 4808 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 24 17:28:48 crc kubenswrapper[4808]: E1124 17:28:48.494825 4808 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:28:49 crc kubenswrapper[4808]: I1124 17:28:49.347072 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:49 crc kubenswrapper[4808]: E1124 17:28:49.347306 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:50 crc kubenswrapper[4808]: I1124 17:28:50.346808 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:50 crc kubenswrapper[4808]: I1124 17:28:50.346912 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:50 crc kubenswrapper[4808]: E1124 17:28:50.347161 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:50 crc kubenswrapper[4808]: E1124 17:28:50.347387 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:50 crc kubenswrapper[4808]: I1124 17:28:50.348557 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:50 crc kubenswrapper[4808]: E1124 17:28:50.348856 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:51 crc kubenswrapper[4808]: I1124 17:28:51.347266 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:51 crc kubenswrapper[4808]: E1124 17:28:51.347769 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:52 crc kubenswrapper[4808]: I1124 17:28:52.346453 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:52 crc kubenswrapper[4808]: I1124 17:28:52.346453 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:52 crc kubenswrapper[4808]: E1124 17:28:52.346604 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:52 crc kubenswrapper[4808]: E1124 17:28:52.346682 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:52 crc kubenswrapper[4808]: I1124 17:28:52.346911 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:52 crc kubenswrapper[4808]: E1124 17:28:52.346996 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:53 crc kubenswrapper[4808]: I1124 17:28:53.347209 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:53 crc kubenswrapper[4808]: E1124 17:28:53.347431 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:53 crc kubenswrapper[4808]: I1124 17:28:53.348743 4808 scope.go:117] "RemoveContainer" containerID="3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222" Nov 24 17:28:53 crc kubenswrapper[4808]: E1124 17:28:53.496638 4808 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:28:54 crc kubenswrapper[4808]: I1124 17:28:54.292253 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-8v2q4"] Nov 24 17:28:54 crc kubenswrapper[4808]: I1124 17:28:54.292457 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:54 crc kubenswrapper[4808]: E1124 17:28:54.292632 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:54 crc kubenswrapper[4808]: I1124 17:28:54.347127 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:54 crc kubenswrapper[4808]: E1124 17:28:54.347876 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:54 crc kubenswrapper[4808]: I1124 17:28:54.347560 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:54 crc kubenswrapper[4808]: E1124 17:28:54.348066 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:54 crc kubenswrapper[4808]: I1124 17:28:54.347493 4808 scope.go:117] "RemoveContainer" containerID="3ac714762891a8496f5f9fb3d368c22acffd162a1a646871d2be0386c78b0a21" Nov 24 17:28:54 crc kubenswrapper[4808]: I1124 17:28:54.368378 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovnkube-controller/3.log" Nov 24 17:28:54 crc kubenswrapper[4808]: I1124 17:28:54.371545 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerStarted","Data":"f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66"} Nov 24 17:28:54 crc kubenswrapper[4808]: I1124 17:28:54.372558 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:28:54 crc kubenswrapper[4808]: I1124 17:28:54.407471 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podStartSLOduration=106.407449669 podStartE2EDuration="1m46.407449669s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:28:54.405523984 +0000 UTC m=+127.003191796" watchObservedRunningTime="2025-11-24 17:28:54.407449669 +0000 UTC m=+127.005117471" Nov 24 17:28:55 crc kubenswrapper[4808]: I1124 17:28:55.346923 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:55 crc kubenswrapper[4808]: E1124 17:28:55.347250 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:55 crc kubenswrapper[4808]: I1124 17:28:55.377715 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q5m49_6a192cd6-853d-4643-855d-ac61168b1cda/kube-multus/1.log" Nov 24 17:28:55 crc kubenswrapper[4808]: I1124 17:28:55.377838 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q5m49" event={"ID":"6a192cd6-853d-4643-855d-ac61168b1cda","Type":"ContainerStarted","Data":"da95fd35f4059c6d375f4d4a2fd7c5e31b6d03712827a2c6b95cdd1678f55c6b"} Nov 24 17:28:56 crc kubenswrapper[4808]: I1124 17:28:56.347251 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:56 crc kubenswrapper[4808]: I1124 17:28:56.347275 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:56 crc kubenswrapper[4808]: I1124 17:28:56.347331 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:56 crc kubenswrapper[4808]: E1124 17:28:56.348076 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:56 crc kubenswrapper[4808]: E1124 17:28:56.348117 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:56 crc kubenswrapper[4808]: E1124 17:28:56.347981 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:57 crc kubenswrapper[4808]: I1124 17:28:57.347312 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:57 crc kubenswrapper[4808]: E1124 17:28:57.347456 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:28:58 crc kubenswrapper[4808]: I1124 17:28:58.346518 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:28:58 crc kubenswrapper[4808]: I1124 17:28:58.346590 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:28:58 crc kubenswrapper[4808]: I1124 17:28:58.348893 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:28:58 crc kubenswrapper[4808]: E1124 17:28:58.348891 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:28:58 crc kubenswrapper[4808]: E1124 17:28:58.349401 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:28:58 crc kubenswrapper[4808]: E1124 17:28:58.349900 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8v2q4" podUID="c26458eb-7f6c-42cb-b6c5-90de6756a994" Nov 24 17:28:59 crc kubenswrapper[4808]: I1124 17:28:59.346729 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:28:59 crc kubenswrapper[4808]: I1124 17:28:59.351698 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 24 17:28:59 crc kubenswrapper[4808]: I1124 17:28:59.351861 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 24 17:29:00 crc kubenswrapper[4808]: I1124 17:29:00.347384 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:29:00 crc kubenswrapper[4808]: I1124 17:29:00.347471 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:29:00 crc kubenswrapper[4808]: I1124 17:29:00.347549 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:29:00 crc kubenswrapper[4808]: I1124 17:29:00.350654 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 24 17:29:00 crc kubenswrapper[4808]: I1124 17:29:00.350691 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 24 17:29:00 crc kubenswrapper[4808]: I1124 17:29:00.351110 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 24 17:29:00 crc kubenswrapper[4808]: I1124 17:29:00.351369 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.618239 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.667159 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.667968 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.668179 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.668771 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.672170 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.673145 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.673652 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.674548 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.674798 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.674805 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.674797 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ns4rm"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.675674 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.676593 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.676773 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.676790 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.676817 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.679673 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.681889 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.681908 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.682466 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.682766 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.683202 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.683255 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.683199 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.683648 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8wq8w"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.684138 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.690334 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qrcr6"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.692388 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.692553 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.704975 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-b5tvs"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.706397 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.707295 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.722903 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.723095 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-z2qxd"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.723592 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dztgx"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.723989 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dztgx" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.724619 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.728235 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vwchc"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.729171 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-77zbq"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.729836 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-6xnwl"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.730707 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-brt82"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.731288 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.731526 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.731840 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.732438 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6xnwl" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.732801 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.733168 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.734004 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.735007 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.742406 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.798945 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.799348 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.799463 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.799632 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.799702 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.799867 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.799901 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.800001 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.798966 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.800479 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.800821 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.801339 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.801586 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.801954 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.802203 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.802404 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.802502 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.802644 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.802873 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.805200 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.805764 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ns4rm"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.805852 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.805870 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.805910 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.805963 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/12cfc041-5c29-43cb-9087-c1cbfb333f8b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-d5bl6\" (UID: \"12cfc041-5c29-43cb-9087-c1cbfb333f8b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.805984 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v94h\" (UniqueName: \"kubernetes.io/projected/47734a14-98f7-4be3-89a3-74f5a13ab3e3-kube-api-access-4v94h\") pod \"downloads-7954f5f757-6xnwl\" (UID: \"47734a14-98f7-4be3-89a3-74f5a13ab3e3\") " pod="openshift-console/downloads-7954f5f757-6xnwl" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806006 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806043 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806062 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806079 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31954b89-9642-4b55-b21e-79882b30b726-serving-cert\") pod \"controller-manager-879f6c89f-vwchc\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806096 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3239fd3e-6afd-4f5f-8a60-a8b338848c8a-config\") pod \"authentication-operator-69f744f599-ns4rm\" (UID: \"3239fd3e-6afd-4f5f-8a60-a8b338848c8a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806113 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn4xw\" (UniqueName: \"kubernetes.io/projected/d54d24ca-c6c4-4313-8f22-908ce8939026-kube-api-access-vn4xw\") pod \"machine-api-operator-5694c8668f-qrcr6\" (UID: \"d54d24ca-c6c4-4313-8f22-908ce8939026\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806136 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a2ec940a-90eb-48d1-82fc-f17260545e94-available-featuregates\") pod \"openshift-config-operator-7777fb866f-77zbq\" (UID: \"a2ec940a-90eb-48d1-82fc-f17260545e94\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806153 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fa927b3-2f38-4d6d-8a79-e25388a60d12-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2dclt\" (UID: \"7fa927b3-2f38-4d6d-8a79-e25388a60d12\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806175 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-node-pullsecrets\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806192 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg5f5\" (UniqueName: \"kubernetes.io/projected/6f47827c-c46c-41e2-8b01-aa13dacca516-kube-api-access-lg5f5\") pod \"machine-approver-56656f9798-wzvvt\" (UID: \"6f47827c-c46c-41e2-8b01-aa13dacca516\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806213 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806231 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vwchc\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806250 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50f0b611-6832-4a5d-8afb-13d5724e8019-client-ca\") pod \"route-controller-manager-6576b87f9c-x5mbp\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806267 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d54d24ca-c6c4-4313-8f22-908ce8939026-config\") pod \"machine-api-operator-5694c8668f-qrcr6\" (UID: \"d54d24ca-c6c4-4313-8f22-908ce8939026\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806286 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f47827c-c46c-41e2-8b01-aa13dacca516-config\") pod \"machine-approver-56656f9798-wzvvt\" (UID: \"6f47827c-c46c-41e2-8b01-aa13dacca516\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806303 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-etcd-client\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806323 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47pf4\" (UniqueName: \"kubernetes.io/projected/aac68768-0e54-4b67-a919-eaa55380171f-kube-api-access-47pf4\") pod \"console-operator-58897d9998-z2qxd\" (UID: \"aac68768-0e54-4b67-a919-eaa55380171f\") " pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806341 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d54d24ca-c6c4-4313-8f22-908ce8939026-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qrcr6\" (UID: \"d54d24ca-c6c4-4313-8f22-908ce8939026\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806360 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806379 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806395 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x8nr\" (UniqueName: \"kubernetes.io/projected/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-kube-api-access-4x8nr\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806422 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fa927b3-2f38-4d6d-8a79-e25388a60d12-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2dclt\" (UID: \"7fa927b3-2f38-4d6d-8a79-e25388a60d12\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806442 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-encryption-config\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806458 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-config\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806475 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1edf7884-088e-497d-a952-76b36a7d867e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-ct2bv\" (UID: \"1edf7884-088e-497d-a952-76b36a7d867e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806493 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50f0b611-6832-4a5d-8afb-13d5724e8019-config\") pod \"route-controller-manager-6576b87f9c-x5mbp\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806509 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806525 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v286k\" (UniqueName: \"kubernetes.io/projected/a2ec940a-90eb-48d1-82fc-f17260545e94-kube-api-access-v286k\") pod \"openshift-config-operator-7777fb866f-77zbq\" (UID: \"a2ec940a-90eb-48d1-82fc-f17260545e94\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806541 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-service-ca\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806556 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1edf7884-088e-497d-a952-76b36a7d867e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-ct2bv\" (UID: \"1edf7884-088e-497d-a952-76b36a7d867e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806573 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzb6v\" (UniqueName: \"kubernetes.io/projected/1edf7884-088e-497d-a952-76b36a7d867e-kube-api-access-lzb6v\") pod \"openshift-controller-manager-operator-756b6f6bc6-ct2bv\" (UID: \"1edf7884-088e-497d-a952-76b36a7d867e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806589 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-config\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806607 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mdcx\" (UniqueName: \"kubernetes.io/projected/31954b89-9642-4b55-b21e-79882b30b726-kube-api-access-7mdcx\") pod \"controller-manager-879f6c89f-vwchc\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806642 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/12cfc041-5c29-43cb-9087-c1cbfb333f8b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-d5bl6\" (UID: \"12cfc041-5c29-43cb-9087-c1cbfb333f8b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806659 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806675 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806691 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-audit\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806716 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgb9v\" (UniqueName: \"kubernetes.io/projected/3239fd3e-6afd-4f5f-8a60-a8b338848c8a-kube-api-access-zgb9v\") pod \"authentication-operator-69f744f599-ns4rm\" (UID: \"3239fd3e-6afd-4f5f-8a60-a8b338848c8a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806736 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/12cfc041-5c29-43cb-9087-c1cbfb333f8b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-d5bl6\" (UID: \"12cfc041-5c29-43cb-9087-c1cbfb333f8b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806753 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806769 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ec940a-90eb-48d1-82fc-f17260545e94-serving-cert\") pod \"openshift-config-operator-7777fb866f-77zbq\" (UID: \"a2ec940a-90eb-48d1-82fc-f17260545e94\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806787 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aac68768-0e54-4b67-a919-eaa55380171f-config\") pod \"console-operator-58897d9998-z2qxd\" (UID: \"aac68768-0e54-4b67-a919-eaa55380171f\") " pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806805 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aac68768-0e54-4b67-a919-eaa55380171f-serving-cert\") pod \"console-operator-58897d9998-z2qxd\" (UID: \"aac68768-0e54-4b67-a919-eaa55380171f\") " pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806822 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgf8c\" (UniqueName: \"kubernetes.io/projected/12cfc041-5c29-43cb-9087-c1cbfb333f8b-kube-api-access-qgf8c\") pod \"cluster-image-registry-operator-dc59b4c8b-d5bl6\" (UID: \"12cfc041-5c29-43cb-9087-c1cbfb333f8b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806839 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-audit-dir\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806855 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-serving-cert\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806874 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m95pc\" (UniqueName: \"kubernetes.io/projected/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-kube-api-access-m95pc\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806891 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/36b5e49a-573b-41ca-aa12-5c991b910a72-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dztgx\" (UID: \"36b5e49a-573b-41ca-aa12-5c991b910a72\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dztgx" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806907 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-audit-dir\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806926 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-config\") pod \"controller-manager-879f6c89f-vwchc\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806943 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6f47827c-c46c-41e2-8b01-aa13dacca516-auth-proxy-config\") pod \"machine-approver-56656f9798-wzvvt\" (UID: \"6f47827c-c46c-41e2-8b01-aa13dacca516\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806961 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3239fd3e-6afd-4f5f-8a60-a8b338848c8a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ns4rm\" (UID: \"3239fd3e-6afd-4f5f-8a60-a8b338848c8a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.806976 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-etcd-serving-ca\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.807002 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-oauth-config\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.807721 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816154 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftdwt\" (UniqueName: \"kubernetes.io/projected/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-kube-api-access-ftdwt\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816223 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-client-ca\") pod \"controller-manager-879f6c89f-vwchc\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816244 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g42d4\" (UniqueName: \"kubernetes.io/projected/7fa927b3-2f38-4d6d-8a79-e25388a60d12-kube-api-access-g42d4\") pod \"openshift-apiserver-operator-796bbdcf4f-2dclt\" (UID: \"7fa927b3-2f38-4d6d-8a79-e25388a60d12\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816270 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3239fd3e-6afd-4f5f-8a60-a8b338848c8a-serving-cert\") pod \"authentication-operator-69f744f599-ns4rm\" (UID: \"3239fd3e-6afd-4f5f-8a60-a8b338848c8a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816287 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-oauth-serving-cert\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816315 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aac68768-0e54-4b67-a919-eaa55380171f-trusted-ca\") pod \"console-operator-58897d9998-z2qxd\" (UID: \"aac68768-0e54-4b67-a919-eaa55380171f\") " pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816333 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d54d24ca-c6c4-4313-8f22-908ce8939026-images\") pod \"machine-api-operator-5694c8668f-qrcr6\" (UID: \"d54d24ca-c6c4-4313-8f22-908ce8939026\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816350 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6f47827c-c46c-41e2-8b01-aa13dacca516-machine-approver-tls\") pod \"machine-approver-56656f9798-wzvvt\" (UID: \"6f47827c-c46c-41e2-8b01-aa13dacca516\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816368 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-serving-cert\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816389 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2mgt\" (UniqueName: \"kubernetes.io/projected/36b5e49a-573b-41ca-aa12-5c991b910a72-kube-api-access-h2mgt\") pod \"cluster-samples-operator-665b6dd947-dztgx\" (UID: \"36b5e49a-573b-41ca-aa12-5c991b910a72\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dztgx" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816410 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-audit-policies\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816426 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3239fd3e-6afd-4f5f-8a60-a8b338848c8a-service-ca-bundle\") pod \"authentication-operator-69f744f599-ns4rm\" (UID: \"3239fd3e-6afd-4f5f-8a60-a8b338848c8a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816445 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50f0b611-6832-4a5d-8afb-13d5724e8019-serving-cert\") pod \"route-controller-manager-6576b87f9c-x5mbp\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816460 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-trusted-ca-bundle\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816481 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdlbv\" (UniqueName: \"kubernetes.io/projected/50f0b611-6832-4a5d-8afb-13d5724e8019-kube-api-access-gdlbv\") pod \"route-controller-manager-6576b87f9c-x5mbp\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.816498 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-image-import-ca\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.823529 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.826104 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.826546 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.826746 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.826961 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.828715 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.829110 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.829279 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.829345 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.829462 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.829581 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.830634 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.829661 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.830830 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.830978 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.829706 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.829719 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.829809 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.829915 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.829948 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.830291 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.831423 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.832661 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-qlpqd"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.833615 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-qlpqd" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.848671 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.851195 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qrcr6"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.856174 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.856218 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.857211 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.858121 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.858162 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.858247 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.858346 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.858360 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.858422 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.858493 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.858561 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.858591 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.858626 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.858668 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.858754 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.858877 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.858969 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.858976 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.859169 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.859350 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.861595 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.875327 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.877894 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.878433 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.883508 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.883778 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.884099 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.884574 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.885105 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.885813 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.886534 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.887832 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.890604 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.918093 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.920228 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.920344 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.921539 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gjnx8"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.922093 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.924314 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.926140 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg5f5\" (UniqueName: \"kubernetes.io/projected/6f47827c-c46c-41e2-8b01-aa13dacca516-kube-api-access-lg5f5\") pod \"machine-approver-56656f9798-wzvvt\" (UID: \"6f47827c-c46c-41e2-8b01-aa13dacca516\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.926177 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.926203 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vwchc\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.926229 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-encryption-config\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.926348 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hmvl\" (UniqueName: \"kubernetes.io/projected/f3a1a3f1-05ca-4f24-9836-cc831fcad234-kube-api-access-2hmvl\") pod \"dns-operator-744455d44c-qlpqd\" (UID: \"f3a1a3f1-05ca-4f24-9836-cc831fcad234\") " pod="openshift-dns-operator/dns-operator-744455d44c-qlpqd" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.948827 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-etcd-client\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949044 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-serving-cert\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949071 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50f0b611-6832-4a5d-8afb-13d5724e8019-client-ca\") pod \"route-controller-manager-6576b87f9c-x5mbp\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949089 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d54d24ca-c6c4-4313-8f22-908ce8939026-config\") pod \"machine-api-operator-5694c8668f-qrcr6\" (UID: \"d54d24ca-c6c4-4313-8f22-908ce8939026\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949109 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f47827c-c46c-41e2-8b01-aa13dacca516-config\") pod \"machine-approver-56656f9798-wzvvt\" (UID: \"6f47827c-c46c-41e2-8b01-aa13dacca516\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949128 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47pf4\" (UniqueName: \"kubernetes.io/projected/aac68768-0e54-4b67-a919-eaa55380171f-kube-api-access-47pf4\") pod \"console-operator-58897d9998-z2qxd\" (UID: \"aac68768-0e54-4b67-a919-eaa55380171f\") " pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949146 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d54d24ca-c6c4-4313-8f22-908ce8939026-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qrcr6\" (UID: \"d54d24ca-c6c4-4313-8f22-908ce8939026\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949162 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949184 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949200 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x8nr\" (UniqueName: \"kubernetes.io/projected/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-kube-api-access-4x8nr\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949224 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949241 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-encryption-config\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949258 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-config\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949325 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1edf7884-088e-497d-a952-76b36a7d867e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-ct2bv\" (UID: \"1edf7884-088e-497d-a952-76b36a7d867e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949368 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fa927b3-2f38-4d6d-8a79-e25388a60d12-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2dclt\" (UID: \"7fa927b3-2f38-4d6d-8a79-e25388a60d12\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949433 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-audit-policies\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949455 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50f0b611-6832-4a5d-8afb-13d5724e8019-config\") pod \"route-controller-manager-6576b87f9c-x5mbp\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949473 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949497 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v286k\" (UniqueName: \"kubernetes.io/projected/a2ec940a-90eb-48d1-82fc-f17260545e94-kube-api-access-v286k\") pod \"openshift-config-operator-7777fb866f-77zbq\" (UID: \"a2ec940a-90eb-48d1-82fc-f17260545e94\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949518 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-service-ca\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949538 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1edf7884-088e-497d-a952-76b36a7d867e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-ct2bv\" (UID: \"1edf7884-088e-497d-a952-76b36a7d867e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949558 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzb6v\" (UniqueName: \"kubernetes.io/projected/1edf7884-088e-497d-a952-76b36a7d867e-kube-api-access-lzb6v\") pod \"openshift-controller-manager-operator-756b6f6bc6-ct2bv\" (UID: \"1edf7884-088e-497d-a952-76b36a7d867e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949580 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-config\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949598 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f3a1a3f1-05ca-4f24-9836-cc831fcad234-metrics-tls\") pod \"dns-operator-744455d44c-qlpqd\" (UID: \"f3a1a3f1-05ca-4f24-9836-cc831fcad234\") " pod="openshift-dns-operator/dns-operator-744455d44c-qlpqd" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949618 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mdcx\" (UniqueName: \"kubernetes.io/projected/31954b89-9642-4b55-b21e-79882b30b726-kube-api-access-7mdcx\") pod \"controller-manager-879f6c89f-vwchc\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949659 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/12cfc041-5c29-43cb-9087-c1cbfb333f8b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-d5bl6\" (UID: \"12cfc041-5c29-43cb-9087-c1cbfb333f8b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949676 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949696 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949715 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-audit\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949746 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgb9v\" (UniqueName: \"kubernetes.io/projected/3239fd3e-6afd-4f5f-8a60-a8b338848c8a-kube-api-access-zgb9v\") pod \"authentication-operator-69f744f599-ns4rm\" (UID: \"3239fd3e-6afd-4f5f-8a60-a8b338848c8a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949766 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/12cfc041-5c29-43cb-9087-c1cbfb333f8b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-d5bl6\" (UID: \"12cfc041-5c29-43cb-9087-c1cbfb333f8b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949784 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949799 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ec940a-90eb-48d1-82fc-f17260545e94-serving-cert\") pod \"openshift-config-operator-7777fb866f-77zbq\" (UID: \"a2ec940a-90eb-48d1-82fc-f17260545e94\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949821 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aac68768-0e54-4b67-a919-eaa55380171f-config\") pod \"console-operator-58897d9998-z2qxd\" (UID: \"aac68768-0e54-4b67-a919-eaa55380171f\") " pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949837 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aac68768-0e54-4b67-a919-eaa55380171f-serving-cert\") pod \"console-operator-58897d9998-z2qxd\" (UID: \"aac68768-0e54-4b67-a919-eaa55380171f\") " pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949854 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgf8c\" (UniqueName: \"kubernetes.io/projected/12cfc041-5c29-43cb-9087-c1cbfb333f8b-kube-api-access-qgf8c\") pod \"cluster-image-registry-operator-dc59b4c8b-d5bl6\" (UID: \"12cfc041-5c29-43cb-9087-c1cbfb333f8b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949870 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m95pc\" (UniqueName: \"kubernetes.io/projected/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-kube-api-access-m95pc\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949890 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-audit-dir\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949906 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-serving-cert\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949923 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/36b5e49a-573b-41ca-aa12-5c991b910a72-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dztgx\" (UID: \"36b5e49a-573b-41ca-aa12-5c991b910a72\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dztgx" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949940 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-audit-dir\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949955 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-config\") pod \"controller-manager-879f6c89f-vwchc\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949976 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-etcd-serving-ca\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.949991 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-oauth-config\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950042 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6f47827c-c46c-41e2-8b01-aa13dacca516-auth-proxy-config\") pod \"machine-approver-56656f9798-wzvvt\" (UID: \"6f47827c-c46c-41e2-8b01-aa13dacca516\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950063 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3239fd3e-6afd-4f5f-8a60-a8b338848c8a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ns4rm\" (UID: \"3239fd3e-6afd-4f5f-8a60-a8b338848c8a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950080 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-client-ca\") pod \"controller-manager-879f6c89f-vwchc\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950097 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g42d4\" (UniqueName: \"kubernetes.io/projected/7fa927b3-2f38-4d6d-8a79-e25388a60d12-kube-api-access-g42d4\") pod \"openshift-apiserver-operator-796bbdcf4f-2dclt\" (UID: \"7fa927b3-2f38-4d6d-8a79-e25388a60d12\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950117 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3239fd3e-6afd-4f5f-8a60-a8b338848c8a-serving-cert\") pod \"authentication-operator-69f744f599-ns4rm\" (UID: \"3239fd3e-6afd-4f5f-8a60-a8b338848c8a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950138 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftdwt\" (UniqueName: \"kubernetes.io/projected/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-kube-api-access-ftdwt\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950155 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-oauth-serving-cert\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950171 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-serving-cert\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950188 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aac68768-0e54-4b67-a919-eaa55380171f-trusted-ca\") pod \"console-operator-58897d9998-z2qxd\" (UID: \"aac68768-0e54-4b67-a919-eaa55380171f\") " pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950205 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d54d24ca-c6c4-4313-8f22-908ce8939026-images\") pod \"machine-api-operator-5694c8668f-qrcr6\" (UID: \"d54d24ca-c6c4-4313-8f22-908ce8939026\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950223 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6f47827c-c46c-41e2-8b01-aa13dacca516-machine-approver-tls\") pod \"machine-approver-56656f9798-wzvvt\" (UID: \"6f47827c-c46c-41e2-8b01-aa13dacca516\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950239 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2mgt\" (UniqueName: \"kubernetes.io/projected/36b5e49a-573b-41ca-aa12-5c991b910a72-kube-api-access-h2mgt\") pod \"cluster-samples-operator-665b6dd947-dztgx\" (UID: \"36b5e49a-573b-41ca-aa12-5c991b910a72\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dztgx" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950256 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-audit-policies\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950271 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3239fd3e-6afd-4f5f-8a60-a8b338848c8a-service-ca-bundle\") pod \"authentication-operator-69f744f599-ns4rm\" (UID: \"3239fd3e-6afd-4f5f-8a60-a8b338848c8a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950287 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950302 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-etcd-client\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950320 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzcq5\" (UniqueName: \"kubernetes.io/projected/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-kube-api-access-qzcq5\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950340 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50f0b611-6832-4a5d-8afb-13d5724e8019-serving-cert\") pod \"route-controller-manager-6576b87f9c-x5mbp\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950357 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-trusted-ca-bundle\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950377 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdlbv\" (UniqueName: \"kubernetes.io/projected/50f0b611-6832-4a5d-8afb-13d5724e8019-kube-api-access-gdlbv\") pod \"route-controller-manager-6576b87f9c-x5mbp\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950412 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-image-import-ca\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950433 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950451 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950470 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v94h\" (UniqueName: \"kubernetes.io/projected/47734a14-98f7-4be3-89a3-74f5a13ab3e3-kube-api-access-4v94h\") pod \"downloads-7954f5f757-6xnwl\" (UID: \"47734a14-98f7-4be3-89a3-74f5a13ab3e3\") " pod="openshift-console/downloads-7954f5f757-6xnwl" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950498 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/12cfc041-5c29-43cb-9087-c1cbfb333f8b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-d5bl6\" (UID: \"12cfc041-5c29-43cb-9087-c1cbfb333f8b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950516 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950534 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950551 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950571 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31954b89-9642-4b55-b21e-79882b30b726-serving-cert\") pod \"controller-manager-879f6c89f-vwchc\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950591 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3239fd3e-6afd-4f5f-8a60-a8b338848c8a-config\") pod \"authentication-operator-69f744f599-ns4rm\" (UID: \"3239fd3e-6afd-4f5f-8a60-a8b338848c8a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950609 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-audit-dir\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950629 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn4xw\" (UniqueName: \"kubernetes.io/projected/d54d24ca-c6c4-4313-8f22-908ce8939026-kube-api-access-vn4xw\") pod \"machine-api-operator-5694c8668f-qrcr6\" (UID: \"d54d24ca-c6c4-4313-8f22-908ce8939026\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.927407 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dztgx"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950867 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-audit-dir\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.950902 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.951837 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a2ec940a-90eb-48d1-82fc-f17260545e94-available-featuregates\") pod \"openshift-config-operator-7777fb866f-77zbq\" (UID: \"a2ec940a-90eb-48d1-82fc-f17260545e94\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.951866 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fa927b3-2f38-4d6d-8a79-e25388a60d12-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2dclt\" (UID: \"7fa927b3-2f38-4d6d-8a79-e25388a60d12\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.951890 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-node-pullsecrets\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.951973 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-node-pullsecrets\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.952091 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.952389 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-brt82"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.952404 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.952452 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-audit-dir\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.952701 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.952716 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.953135 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-client-ca\") pod \"controller-manager-879f6c89f-vwchc\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.953167 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-etcd-serving-ca\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.954549 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-config\") pod \"controller-manager-879f6c89f-vwchc\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.961826 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3239fd3e-6afd-4f5f-8a60-a8b338848c8a-service-ca-bundle\") pod \"authentication-operator-69f744f599-ns4rm\" (UID: \"3239fd3e-6afd-4f5f-8a60-a8b338848c8a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.962474 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vwchc\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.962797 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-serving-cert\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.962978 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/36b5e49a-573b-41ca-aa12-5c991b910a72-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dztgx\" (UID: \"36b5e49a-573b-41ca-aa12-5c991b910a72\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dztgx" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.964144 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-audit-policies\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.965050 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.971760 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.973899 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50f0b611-6832-4a5d-8afb-13d5724e8019-serving-cert\") pod \"route-controller-manager-6576b87f9c-x5mbp\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.974997 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.977317 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3239fd3e-6afd-4f5f-8a60-a8b338848c8a-serving-cert\") pod \"authentication-operator-69f744f599-ns4rm\" (UID: \"3239fd3e-6afd-4f5f-8a60-a8b338848c8a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.978127 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6f47827c-c46c-41e2-8b01-aa13dacca516-auth-proxy-config\") pod \"machine-approver-56656f9798-wzvvt\" (UID: \"6f47827c-c46c-41e2-8b01-aa13dacca516\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.979889 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d54d24ca-c6c4-4313-8f22-908ce8939026-images\") pod \"machine-api-operator-5694c8668f-qrcr6\" (UID: \"d54d24ca-c6c4-4313-8f22-908ce8939026\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.980410 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.980434 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a2ec940a-90eb-48d1-82fc-f17260545e94-available-featuregates\") pod \"openshift-config-operator-7777fb866f-77zbq\" (UID: \"a2ec940a-90eb-48d1-82fc-f17260545e94\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.980487 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vwchc"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.992353 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.992386 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-mmrgn"] Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.992693 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f47827c-c46c-41e2-8b01-aa13dacca516-config\") pod \"machine-approver-56656f9798-wzvvt\" (UID: \"6f47827c-c46c-41e2-8b01-aa13dacca516\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.984293 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aac68768-0e54-4b67-a919-eaa55380171f-trusted-ca\") pod \"console-operator-58897d9998-z2qxd\" (UID: \"aac68768-0e54-4b67-a919-eaa55380171f\") " pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.984762 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-oauth-serving-cert\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.929689 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.941334 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.985473 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-trusted-ca-bundle\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.987125 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-image-import-ca\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.987178 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-service-ca\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.989676 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.989750 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.990093 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50f0b611-6832-4a5d-8afb-13d5724e8019-client-ca\") pod \"route-controller-manager-6576b87f9c-x5mbp\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.991031 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3239fd3e-6afd-4f5f-8a60-a8b338848c8a-config\") pod \"authentication-operator-69f744f599-ns4rm\" (UID: \"3239fd3e-6afd-4f5f-8a60-a8b338848c8a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.991444 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-etcd-client\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.991995 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d54d24ca-c6c4-4313-8f22-908ce8939026-config\") pod \"machine-api-operator-5694c8668f-qrcr6\" (UID: \"d54d24ca-c6c4-4313-8f22-908ce8939026\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.934496 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.982860 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3239fd3e-6afd-4f5f-8a60-a8b338848c8a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ns4rm\" (UID: \"3239fd3e-6afd-4f5f-8a60-a8b338848c8a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.934694 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.935141 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.940267 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.941074 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.941996 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.981170 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-oauth-config\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.981884 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" Nov 24 17:29:07 crc kubenswrapper[4808]: I1124 17:29:07.942975 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.943562 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.944402 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.995291 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.945413 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.947320 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.948660 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.951158 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.996002 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.996623 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fa927b3-2f38-4d6d-8a79-e25388a60d12-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2dclt\" (UID: \"7fa927b3-2f38-4d6d-8a79-e25388a60d12\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.996817 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/12cfc041-5c29-43cb-9087-c1cbfb333f8b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-d5bl6\" (UID: \"12cfc041-5c29-43cb-9087-c1cbfb333f8b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.997033 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.997337 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1edf7884-088e-497d-a952-76b36a7d867e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-ct2bv\" (UID: \"1edf7884-088e-497d-a952-76b36a7d867e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.997548 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-serving-cert\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.997820 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.997878 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50f0b611-6832-4a5d-8afb-13d5724e8019-config\") pod \"route-controller-manager-6576b87f9c-x5mbp\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.997926 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/12cfc041-5c29-43cb-9087-c1cbfb333f8b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-d5bl6\" (UID: \"12cfc041-5c29-43cb-9087-c1cbfb333f8b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.998092 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.998293 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-audit\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.998402 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ec940a-90eb-48d1-82fc-f17260545e94-serving-cert\") pod \"openshift-config-operator-7777fb866f-77zbq\" (UID: \"a2ec940a-90eb-48d1-82fc-f17260545e94\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.998750 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6f47827c-c46c-41e2-8b01-aa13dacca516-machine-approver-tls\") pod \"machine-approver-56656f9798-wzvvt\" (UID: \"6f47827c-c46c-41e2-8b01-aa13dacca516\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:07.999646 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fa927b3-2f38-4d6d-8a79-e25388a60d12-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2dclt\" (UID: \"7fa927b3-2f38-4d6d-8a79-e25388a60d12\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.000562 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aac68768-0e54-4b67-a919-eaa55380171f-config\") pod \"console-operator-58897d9998-z2qxd\" (UID: \"aac68768-0e54-4b67-a919-eaa55380171f\") " pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.000911 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.001552 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aac68768-0e54-4b67-a919-eaa55380171f-serving-cert\") pod \"console-operator-58897d9998-z2qxd\" (UID: \"aac68768-0e54-4b67-a919-eaa55380171f\") " pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.002475 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d54d24ca-c6c4-4313-8f22-908ce8939026-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qrcr6\" (UID: \"d54d24ca-c6c4-4313-8f22-908ce8939026\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.005564 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-config\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.005937 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.006000 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-encryption-config\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.006154 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.006493 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-config\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.006744 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31954b89-9642-4b55-b21e-79882b30b726-serving-cert\") pod \"controller-manager-879f6c89f-vwchc\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.007487 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vkxbx"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.007681 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.008486 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.008739 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.009173 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.009655 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5c5r4"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.010106 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.010219 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1edf7884-088e-497d-a952-76b36a7d867e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-ct2bv\" (UID: \"1edf7884-088e-497d-a952-76b36a7d867e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.010340 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.010594 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.010858 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5c5r4" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.010896 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.012643 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ljn6r"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.013643 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ljn6r" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.014214 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.020827 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.021424 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-b5tvs"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.022644 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6xnwl"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.023667 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.024230 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.026164 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2v2lb"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.027337 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.027554 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8wq8w"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.028733 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-nvgpx"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.029620 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nvgpx" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.030210 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.030226 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.031410 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.031808 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.032418 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.036420 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.037919 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-z2qxd"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.038090 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.039044 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.039780 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.040291 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.041667 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-qlpqd"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.042722 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.044120 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tls4q"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.045569 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-2d6vb"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.045755 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-tls4q" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.046164 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-2d6vb" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.047848 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.049386 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.050207 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.050431 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wvj5k"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.051597 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.052255 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.052573 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.052615 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-audit-policies\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.052668 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f3a1a3f1-05ca-4f24-9836-cc831fcad234-metrics-tls\") pod \"dns-operator-744455d44c-qlpqd\" (UID: \"f3a1a3f1-05ca-4f24-9836-cc831fcad234\") " pod="openshift-dns-operator/dns-operator-744455d44c-qlpqd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.052784 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9c6772a-c49e-441b-86a3-b76d1ea62823-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-chwdd\" (UID: \"e9c6772a-c49e-441b-86a3-b76d1ea62823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.052811 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.052836 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-etcd-client\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.052857 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzcq5\" (UniqueName: \"kubernetes.io/projected/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-kube-api-access-qzcq5\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.052885 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c6772a-c49e-441b-86a3-b76d1ea62823-config\") pod \"kube-apiserver-operator-766d6c64bb-chwdd\" (UID: \"e9c6772a-c49e-441b-86a3-b76d1ea62823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.052902 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9c6772a-c49e-441b-86a3-b76d1ea62823-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-chwdd\" (UID: \"e9c6772a-c49e-441b-86a3-b76d1ea62823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.052935 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-audit-dir\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.052966 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hmvl\" (UniqueName: \"kubernetes.io/projected/f3a1a3f1-05ca-4f24-9836-cc831fcad234-kube-api-access-2hmvl\") pod \"dns-operator-744455d44c-qlpqd\" (UID: \"f3a1a3f1-05ca-4f24-9836-cc831fcad234\") " pod="openshift-dns-operator/dns-operator-744455d44c-qlpqd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.052988 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-encryption-config\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.053063 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-audit-dir\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.053188 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-serving-cert\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.053235 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.053525 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gjnx8"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.053741 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-audit-policies\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.054101 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.054725 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vkxbx"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.056154 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5c5r4"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.056748 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-encryption-config\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.056818 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-serving-cert\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.057135 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2v2lb"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.057287 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-etcd-client\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.057385 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f3a1a3f1-05ca-4f24-9836-cc831fcad234-metrics-tls\") pod \"dns-operator-744455d44c-qlpqd\" (UID: \"f3a1a3f1-05ca-4f24-9836-cc831fcad234\") " pod="openshift-dns-operator/dns-operator-744455d44c-qlpqd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.058322 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.059285 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.060502 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-77zbq"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.061506 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.062657 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ljn6r"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.064547 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-nvgpx"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.065997 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.068009 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.069669 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.071604 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.072770 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.077227 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.083997 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tls4q"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.085866 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-2d6vb"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.089712 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.090862 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wvj5k"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.092445 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-7sbnn"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.093339 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-m57fw"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.093377 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7sbnn" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.093984 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-m57fw" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.094655 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg5f5\" (UniqueName: \"kubernetes.io/projected/6f47827c-c46c-41e2-8b01-aa13dacca516-kube-api-access-lg5f5\") pod \"machine-approver-56656f9798-wzvvt\" (UID: \"6f47827c-c46c-41e2-8b01-aa13dacca516\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.096318 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-m57fw"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.125503 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2mgt\" (UniqueName: \"kubernetes.io/projected/36b5e49a-573b-41ca-aa12-5c991b910a72-kube-api-access-h2mgt\") pod \"cluster-samples-operator-665b6dd947-dztgx\" (UID: \"36b5e49a-573b-41ca-aa12-5c991b910a72\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dztgx" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.145233 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgf8c\" (UniqueName: \"kubernetes.io/projected/12cfc041-5c29-43cb-9087-c1cbfb333f8b-kube-api-access-qgf8c\") pod \"cluster-image-registry-operator-dc59b4c8b-d5bl6\" (UID: \"12cfc041-5c29-43cb-9087-c1cbfb333f8b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.154262 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9c6772a-c49e-441b-86a3-b76d1ea62823-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-chwdd\" (UID: \"e9c6772a-c49e-441b-86a3-b76d1ea62823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.154317 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c6772a-c49e-441b-86a3-b76d1ea62823-config\") pod \"kube-apiserver-operator-766d6c64bb-chwdd\" (UID: \"e9c6772a-c49e-441b-86a3-b76d1ea62823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.154342 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9c6772a-c49e-441b-86a3-b76d1ea62823-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-chwdd\" (UID: \"e9c6772a-c49e-441b-86a3-b76d1ea62823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.156054 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c6772a-c49e-441b-86a3-b76d1ea62823-config\") pod \"kube-apiserver-operator-766d6c64bb-chwdd\" (UID: \"e9c6772a-c49e-441b-86a3-b76d1ea62823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.158095 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9c6772a-c49e-441b-86a3-b76d1ea62823-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-chwdd\" (UID: \"e9c6772a-c49e-441b-86a3-b76d1ea62823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.165705 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m95pc\" (UniqueName: \"kubernetes.io/projected/8b35999c-bfc9-44bd-bca5-a97e645a7ed9-kube-api-access-m95pc\") pod \"apiserver-76f77b778f-8wq8w\" (UID: \"8b35999c-bfc9-44bd-bca5-a97e645a7ed9\") " pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.185088 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g42d4\" (UniqueName: \"kubernetes.io/projected/7fa927b3-2f38-4d6d-8a79-e25388a60d12-kube-api-access-g42d4\") pod \"openshift-apiserver-operator-796bbdcf4f-2dclt\" (UID: \"7fa927b3-2f38-4d6d-8a79-e25388a60d12\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.205106 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v94h\" (UniqueName: \"kubernetes.io/projected/47734a14-98f7-4be3-89a3-74f5a13ab3e3-kube-api-access-4v94h\") pod \"downloads-7954f5f757-6xnwl\" (UID: \"47734a14-98f7-4be3-89a3-74f5a13ab3e3\") " pod="openshift-console/downloads-7954f5f757-6xnwl" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.226789 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdlbv\" (UniqueName: \"kubernetes.io/projected/50f0b611-6832-4a5d-8afb-13d5724e8019-kube-api-access-gdlbv\") pod \"route-controller-manager-6576b87f9c-x5mbp\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.248271 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftdwt\" (UniqueName: \"kubernetes.io/projected/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-kube-api-access-ftdwt\") pod \"oauth-openshift-558db77b4-b5tvs\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.250246 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.270655 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.290451 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.297797 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.311309 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 24 17:29:08 crc kubenswrapper[4808]: W1124 17:29:08.316445 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f47827c_c46c_41e2_8b01_aa13dacca516.slice/crio-07cc6d302e1b71191dca15deb017cfa0769dccb46417c982e0d8b573acd10fa0 WatchSource:0}: Error finding container 07cc6d302e1b71191dca15deb017cfa0769dccb46417c982e0d8b573acd10fa0: Status 404 returned error can't find the container with id 07cc6d302e1b71191dca15deb017cfa0769dccb46417c982e0d8b573acd10fa0 Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.319361 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.335326 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.347393 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.365148 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v286k\" (UniqueName: \"kubernetes.io/projected/a2ec940a-90eb-48d1-82fc-f17260545e94-kube-api-access-v286k\") pod \"openshift-config-operator-7777fb866f-77zbq\" (UID: \"a2ec940a-90eb-48d1-82fc-f17260545e94\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.389090 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn4xw\" (UniqueName: \"kubernetes.io/projected/d54d24ca-c6c4-4313-8f22-908ce8939026-kube-api-access-vn4xw\") pod \"machine-api-operator-5694c8668f-qrcr6\" (UID: \"d54d24ca-c6c4-4313-8f22-908ce8939026\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.400388 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.405361 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgb9v\" (UniqueName: \"kubernetes.io/projected/3239fd3e-6afd-4f5f-8a60-a8b338848c8a-kube-api-access-zgb9v\") pod \"authentication-operator-69f744f599-ns4rm\" (UID: \"3239fd3e-6afd-4f5f-8a60-a8b338848c8a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.405756 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.418320 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dztgx" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.427895 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47pf4\" (UniqueName: \"kubernetes.io/projected/aac68768-0e54-4b67-a919-eaa55380171f-kube-api-access-47pf4\") pod \"console-operator-58897d9998-z2qxd\" (UID: \"aac68768-0e54-4b67-a919-eaa55380171f\") " pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.430419 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" event={"ID":"6f47827c-c46c-41e2-8b01-aa13dacca516","Type":"ContainerStarted","Data":"07cc6d302e1b71191dca15deb017cfa0769dccb46417c982e0d8b573acd10fa0"} Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.433417 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.437858 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.459124 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6xnwl" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.470008 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.470602 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzb6v\" (UniqueName: \"kubernetes.io/projected/1edf7884-088e-497d-a952-76b36a7d867e-kube-api-access-lzb6v\") pod \"openshift-controller-manager-operator-756b6f6bc6-ct2bv\" (UID: \"1edf7884-088e-497d-a952-76b36a7d867e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.490522 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.514783 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.530856 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.554295 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.571686 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.585165 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.612122 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.615982 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/12cfc041-5c29-43cb-9087-c1cbfb333f8b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-d5bl6\" (UID: \"12cfc041-5c29-43cb-9087-c1cbfb333f8b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" Nov 24 17:29:08 crc kubenswrapper[4808]: W1124 17:29:08.624613 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fa927b3_2f38_4d6d_8a79_e25388a60d12.slice/crio-f550fb4fa5e64b890063a33184a3cdc47a18ddb346128fbe1d4057769b891645 WatchSource:0}: Error finding container f550fb4fa5e64b890063a33184a3cdc47a18ddb346128fbe1d4057769b891645: Status 404 returned error can't find the container with id f550fb4fa5e64b890063a33184a3cdc47a18ddb346128fbe1d4057769b891645 Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.627065 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.631500 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.638011 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.659993 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" Nov 24 17:29:08 crc kubenswrapper[4808]: W1124 17:29:08.671429 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0416a1dd_4c58_4d18_b6f3_404e4c642e6b.slice/crio-1a81f64f0b332f7acf2ed04c8759611e350f37d94a302dfd8742848c16388436 WatchSource:0}: Error finding container 1a81f64f0b332f7acf2ed04c8759611e350f37d94a302dfd8742848c16388436: Status 404 returned error can't find the container with id 1a81f64f0b332f7acf2ed04c8759611e350f37d94a302dfd8742848c16388436 Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.672360 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.672624 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-b5tvs"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.672847 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x8nr\" (UniqueName: \"kubernetes.io/projected/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-kube-api-access-4x8nr\") pod \"console-f9d7485db-brt82\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.672973 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.712793 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mdcx\" (UniqueName: \"kubernetes.io/projected/31954b89-9642-4b55-b21e-79882b30b726-kube-api-access-7mdcx\") pod \"controller-manager-879f6c89f-vwchc\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.719552 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8wq8w"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.733373 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 24 17:29:08 crc kubenswrapper[4808]: W1124 17:29:08.737098 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b35999c_bfc9_44bd_bca5_a97e645a7ed9.slice/crio-8ed75d94f497411bcc3b379fc0a378a6bbbf6af50aa01ef3c61752ddef093fd1 WatchSource:0}: Error finding container 8ed75d94f497411bcc3b379fc0a378a6bbbf6af50aa01ef3c61752ddef093fd1: Status 404 returned error can't find the container with id 8ed75d94f497411bcc3b379fc0a378a6bbbf6af50aa01ef3c61752ddef093fd1 Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.744040 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.751965 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.758707 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.770717 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.791505 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.811680 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.816214 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.835414 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.836840 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.851142 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.870606 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.891250 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.907594 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-77zbq"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.911056 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.929836 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ns4rm"] Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.930635 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.937549 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.955426 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.963661 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qrcr6"] Nov 24 17:29:08 crc kubenswrapper[4808]: W1124 17:29:08.969396 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3239fd3e_6afd_4f5f_8a60_a8b338848c8a.slice/crio-6b26b3894738bc8f049d089d099343be322ce256934f0747e427d95188095173 WatchSource:0}: Error finding container 6b26b3894738bc8f049d089d099343be322ce256934f0747e427d95188095173: Status 404 returned error can't find the container with id 6b26b3894738bc8f049d089d099343be322ce256934f0747e427d95188095173 Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.971388 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 24 17:29:08 crc kubenswrapper[4808]: W1124 17:29:08.984509 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd54d24ca_c6c4_4313_8f22_908ce8939026.slice/crio-e9123db7f7bae8be53fecd33d0ea790a10ba383d5b7c79365e4c62e472e06741 WatchSource:0}: Error finding container e9123db7f7bae8be53fecd33d0ea790a10ba383d5b7c79365e4c62e472e06741: Status 404 returned error can't find the container with id e9123db7f7bae8be53fecd33d0ea790a10ba383d5b7c79365e4c62e472e06741 Nov 24 17:29:08 crc kubenswrapper[4808]: I1124 17:29:08.991087 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.008503 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dztgx"] Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.013210 4808 request.go:700] Waited for 1.003821782s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-etcd-operator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.016563 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.018727 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-z2qxd"] Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.031776 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.031774 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6xnwl"] Nov 24 17:29:09 crc kubenswrapper[4808]: W1124 17:29:09.046620 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaac68768_0e54_4b67_a919_eaa55380171f.slice/crio-8bde5f01e4043f5bf96f5415af00d5554340334ae6c07058ef4b07e45381c18d WatchSource:0}: Error finding container 8bde5f01e4043f5bf96f5415af00d5554340334ae6c07058ef4b07e45381c18d: Status 404 returned error can't find the container with id 8bde5f01e4043f5bf96f5415af00d5554340334ae6c07058ef4b07e45381c18d Nov 24 17:29:09 crc kubenswrapper[4808]: W1124 17:29:09.046855 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47734a14_98f7_4be3_89a3_74f5a13ab3e3.slice/crio-2156eae67e3ca09c04e6bc44a1dae71c80debec8accb1a44acf6e9f88e9e31c6 WatchSource:0}: Error finding container 2156eae67e3ca09c04e6bc44a1dae71c80debec8accb1a44acf6e9f88e9e31c6: Status 404 returned error can't find the container with id 2156eae67e3ca09c04e6bc44a1dae71c80debec8accb1a44acf6e9f88e9e31c6 Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.050643 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.072569 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.083332 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv"] Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.094182 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.110941 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.133457 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.150641 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-brt82"] Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.151163 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.171565 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 24 17:29:09 crc kubenswrapper[4808]: W1124 17:29:09.179642 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1edf7884_088e_497d_a952_76b36a7d867e.slice/crio-77ca94f9aecff10a2b3f65d926bc1832a162baaad788f4ed89b1134ec09822d7 WatchSource:0}: Error finding container 77ca94f9aecff10a2b3f65d926bc1832a162baaad788f4ed89b1134ec09822d7: Status 404 returned error can't find the container with id 77ca94f9aecff10a2b3f65d926bc1832a162baaad788f4ed89b1134ec09822d7 Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.181718 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6"] Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.191675 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.220809 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.231405 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.244286 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vwchc"] Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.250120 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.270590 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.290825 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.310298 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.330800 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 24 17:29:09 crc kubenswrapper[4808]: W1124 17:29:09.338596 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31954b89_9642_4b55_b21e_79882b30b726.slice/crio-aa8eb98c2ebd24ee01f789edf13676f8b49c05fcea68b911bb71dbedb78371fb WatchSource:0}: Error finding container aa8eb98c2ebd24ee01f789edf13676f8b49c05fcea68b911bb71dbedb78371fb: Status 404 returned error can't find the container with id aa8eb98c2ebd24ee01f789edf13676f8b49c05fcea68b911bb71dbedb78371fb Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.349706 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.371451 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.390691 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.410467 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.430929 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.438182 4808 generic.go:334] "Generic (PLEG): container finished" podID="a2ec940a-90eb-48d1-82fc-f17260545e94" containerID="d75d1b638d9ac724d3f8828c1a22d2cdbd2758b87f11c9c7261845d20f716218" exitCode=0 Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.438297 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" event={"ID":"a2ec940a-90eb-48d1-82fc-f17260545e94","Type":"ContainerDied","Data":"d75d1b638d9ac724d3f8828c1a22d2cdbd2758b87f11c9c7261845d20f716218"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.438335 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" event={"ID":"a2ec940a-90eb-48d1-82fc-f17260545e94","Type":"ContainerStarted","Data":"7f894b356186a11669e15cff0aef9f3e6c8ff2abc66b0bcc3789f29527c497d1"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.439949 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-brt82" event={"ID":"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d","Type":"ContainerStarted","Data":"b747ed7262c0d73c9e8deb41ab48fcd9b1133c1d57bd14c438347f06773b9199"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.441289 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" event={"ID":"50f0b611-6832-4a5d-8afb-13d5724e8019","Type":"ContainerStarted","Data":"1a216ed389533603c08fb010bd19f359db30e13e36256973b9cce788bd7d8a80"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.441319 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" event={"ID":"50f0b611-6832-4a5d-8afb-13d5724e8019","Type":"ContainerStarted","Data":"c9e8f19112e3df39d36bc1fa35a9a22f6691de6466e4d480b1d90d7133e205b2"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.442092 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.443537 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv" event={"ID":"1edf7884-088e-497d-a952-76b36a7d867e","Type":"ContainerStarted","Data":"77ca94f9aecff10a2b3f65d926bc1832a162baaad788f4ed89b1134ec09822d7"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.443892 4808 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-x5mbp container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.443934 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" podUID="50f0b611-6832-4a5d-8afb-13d5724e8019" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.445616 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dztgx" event={"ID":"36b5e49a-573b-41ca-aa12-5c991b910a72","Type":"ContainerStarted","Data":"aaab48fb936f175f429fa3aab26f8f1912ed6324a10fba6eb996c21d1adb2804"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.451148 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.451584 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-z2qxd" event={"ID":"aac68768-0e54-4b67-a919-eaa55380171f","Type":"ContainerStarted","Data":"f52bb5fc38d8733ff248fddd64f3f9c2b6107157c7da917b35504b8dab7b98d2"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.451932 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-z2qxd" event={"ID":"aac68768-0e54-4b67-a919-eaa55380171f","Type":"ContainerStarted","Data":"8bde5f01e4043f5bf96f5415af00d5554340334ae6c07058ef4b07e45381c18d"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.451975 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.457404 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" event={"ID":"31954b89-9642-4b55-b21e-79882b30b726","Type":"ContainerStarted","Data":"aa8eb98c2ebd24ee01f789edf13676f8b49c05fcea68b911bb71dbedb78371fb"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.457725 4808 patch_prober.go:28] interesting pod/console-operator-58897d9998-z2qxd container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.457799 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-z2qxd" podUID="aac68768-0e54-4b67-a919-eaa55380171f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.459672 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" event={"ID":"12cfc041-5c29-43cb-9087-c1cbfb333f8b","Type":"ContainerStarted","Data":"a51ef4e82fbf3b500f9e38b43b7142fac101d9f1270bc9e6a81cb0c53d487f4b"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.476315 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.481438 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" event={"ID":"3239fd3e-6afd-4f5f-8a60-a8b338848c8a","Type":"ContainerStarted","Data":"3f1a4f0d28b1838ad4eb9f50181be8e82b414352a296080bd52a2512baf7d282"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.481481 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" event={"ID":"3239fd3e-6afd-4f5f-8a60-a8b338848c8a","Type":"ContainerStarted","Data":"6b26b3894738bc8f049d089d099343be322ce256934f0747e427d95188095173"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.489083 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" event={"ID":"6f47827c-c46c-41e2-8b01-aa13dacca516","Type":"ContainerStarted","Data":"b66560dd398456985b987c3d1792c5164c063c982f89d32a69f719d61b3fb1f2"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.489136 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" event={"ID":"6f47827c-c46c-41e2-8b01-aa13dacca516","Type":"ContainerStarted","Data":"474bbbd1bbcb3ba51fdb5d235e0c4bb773e1afa2f5fa4a9a9cf8285a6a623aba"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.490733 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.492126 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt" event={"ID":"7fa927b3-2f38-4d6d-8a79-e25388a60d12","Type":"ContainerStarted","Data":"5cf52b4fa1c1f9a9a988c0f01aaa43c1700551cc28338e3b69bba91e34ec0054"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.492168 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt" event={"ID":"7fa927b3-2f38-4d6d-8a79-e25388a60d12","Type":"ContainerStarted","Data":"f550fb4fa5e64b890063a33184a3cdc47a18ddb346128fbe1d4057769b891645"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.495333 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" event={"ID":"0416a1dd-4c58-4d18-b6f3-404e4c642e6b","Type":"ContainerStarted","Data":"ba947ce528c834968431673cff84469bc626c08c90654a64e43ad10b0953a273"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.495413 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" event={"ID":"0416a1dd-4c58-4d18-b6f3-404e4c642e6b","Type":"ContainerStarted","Data":"1a81f64f0b332f7acf2ed04c8759611e350f37d94a302dfd8742848c16388436"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.495604 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.497836 4808 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-b5tvs container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.9:6443/healthz\": dial tcp 10.217.0.9:6443: connect: connection refused" start-of-body= Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.497937 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6xnwl" event={"ID":"47734a14-98f7-4be3-89a3-74f5a13ab3e3","Type":"ContainerStarted","Data":"ebd7298072e8ee786814450d66461d84b282c66e8dfd56739f17fc1a870bc0de"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.497986 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6xnwl" event={"ID":"47734a14-98f7-4be3-89a3-74f5a13ab3e3","Type":"ContainerStarted","Data":"2156eae67e3ca09c04e6bc44a1dae71c80debec8accb1a44acf6e9f88e9e31c6"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.498160 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6xnwl" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.497875 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" podUID="0416a1dd-4c58-4d18-b6f3-404e4c642e6b" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.9:6443/healthz\": dial tcp 10.217.0.9:6443: connect: connection refused" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.499544 4808 generic.go:334] "Generic (PLEG): container finished" podID="8b35999c-bfc9-44bd-bca5-a97e645a7ed9" containerID="2da0ee239f624e5ab3fdc19683c93187e14fc2c570dfd8c4301f830bdda22482" exitCode=0 Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.499592 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" event={"ID":"8b35999c-bfc9-44bd-bca5-a97e645a7ed9","Type":"ContainerDied","Data":"2da0ee239f624e5ab3fdc19683c93187e14fc2c570dfd8c4301f830bdda22482"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.499607 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" event={"ID":"8b35999c-bfc9-44bd-bca5-a97e645a7ed9","Type":"ContainerStarted","Data":"8ed75d94f497411bcc3b379fc0a378a6bbbf6af50aa01ef3c61752ddef093fd1"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.500293 4808 patch_prober.go:28] interesting pod/downloads-7954f5f757-6xnwl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.500364 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6xnwl" podUID="47734a14-98f7-4be3-89a3-74f5a13ab3e3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.504119 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" event={"ID":"d54d24ca-c6c4-4313-8f22-908ce8939026","Type":"ContainerStarted","Data":"d81f1439a3128ea691db8b708b77735c152026f4b6dba76618ddb9d1232e28ba"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.504173 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" event={"ID":"d54d24ca-c6c4-4313-8f22-908ce8939026","Type":"ContainerStarted","Data":"e9123db7f7bae8be53fecd33d0ea790a10ba383d5b7c79365e4c62e472e06741"} Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.512772 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.531624 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.551049 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.571339 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.591197 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.611865 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.630949 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.651091 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.670584 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.691667 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.710943 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.732143 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.751195 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.776222 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.790856 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.811285 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.831138 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.852043 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.870294 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.891116 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.912468 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.930796 4808 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.968817 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzcq5\" (UniqueName: \"kubernetes.io/projected/59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c-kube-api-access-qzcq5\") pod \"apiserver-7bbb656c7d-7bhtk\" (UID: \"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:09 crc kubenswrapper[4808]: I1124 17:29:09.990230 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hmvl\" (UniqueName: \"kubernetes.io/projected/f3a1a3f1-05ca-4f24-9836-cc831fcad234-kube-api-access-2hmvl\") pod \"dns-operator-744455d44c-qlpqd\" (UID: \"f3a1a3f1-05ca-4f24-9836-cc831fcad234\") " pod="openshift-dns-operator/dns-operator-744455d44c-qlpqd" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.003779 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/edb7cea2-f6eb-4043-91d9-c7abaef74455-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.004343 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/edb7cea2-f6eb-4043-91d9-c7abaef74455-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.004509 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-registry-tls\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.004723 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/edb7cea2-f6eb-4043-91d9-c7abaef74455-trusted-ca\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.004902 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/edb7cea2-f6eb-4043-91d9-c7abaef74455-registry-certificates\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.005282 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.005405 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqbq8\" (UniqueName: \"kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-kube-api-access-mqbq8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.005800 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-bound-sa-token\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: E1124 17:29:10.007690 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:10.507664613 +0000 UTC m=+143.105332415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.013533 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.029597 4808 request.go:700] Waited for 1.935894308s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dnode-bootstrapper-token&limit=500&resourceVersion=0 Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.032168 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.052304 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.070058 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.091751 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.106808 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:10 crc kubenswrapper[4808]: E1124 17:29:10.107106 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:10.607061993 +0000 UTC m=+143.204729835 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109065 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7plk6\" (UniqueName: \"kubernetes.io/projected/4319fe26-e7cc-4665-881a-6275268b6330-kube-api-access-7plk6\") pod \"control-plane-machine-set-operator-78cbb6b69f-ljn6r\" (UID: \"4319fe26-e7cc-4665-881a-6275268b6330\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ljn6r" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109133 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df575f42-a7d3-42d8-9d03-f5334bbbc35b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfd6f\" (UID: \"df575f42-a7d3-42d8-9d03-f5334bbbc35b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109186 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29292c32-1a21-415c-83ce-1f526a0264b5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kgrtc\" (UID: \"29292c32-1a21-415c-83ce-1f526a0264b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109234 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/22764299-441e-4ac3-986a-5a4c1a325816-proxy-tls\") pod \"machine-config-operator-74547568cd-c4svp\" (UID: \"22764299-441e-4ac3-986a-5a4c1a325816\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109334 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a54581b-46f9-4453-b3c0-9cb55e3410f6-service-ca-bundle\") pod \"router-default-5444994796-mmrgn\" (UID: \"3a54581b-46f9-4453-b3c0-9cb55e3410f6\") " pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109366 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68fp2\" (UniqueName: \"kubernetes.io/projected/4c27c616-6ab4-40fb-857f-cfd4d507547e-kube-api-access-68fp2\") pod \"olm-operator-6b444d44fb-xlpkq\" (UID: \"4c27c616-6ab4-40fb-857f-cfd4d507547e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109530 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqbq8\" (UniqueName: \"kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-kube-api-access-mqbq8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109626 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlzm7\" (UniqueName: \"kubernetes.io/projected/72380442-6dd0-45fd-82ce-07191fb9427e-kube-api-access-nlzm7\") pod \"kube-storage-version-migrator-operator-b67b599dd-w2bpk\" (UID: \"72380442-6dd0-45fd-82ce-07191fb9427e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109699 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0706d14f-f31b-4935-a1f5-6e1af9f11b13-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zcmdg\" (UID: \"0706d14f-f31b-4935-a1f5-6e1af9f11b13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109733 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m75ks\" (UniqueName: \"kubernetes.io/projected/81133ff7-9de8-4958-b47c-7513df033b38-kube-api-access-m75ks\") pod \"machine-config-controller-84d6567774-9dqfr\" (UID: \"81133ff7-9de8-4958-b47c-7513df033b38\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109796 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-serving-cert\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109841 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0706d14f-f31b-4935-a1f5-6e1af9f11b13-metrics-tls\") pod \"ingress-operator-5b745b69d9-zcmdg\" (UID: \"0706d14f-f31b-4935-a1f5-6e1af9f11b13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109874 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v28m2\" (UniqueName: \"kubernetes.io/projected/1b489c90-5e8b-4a49-8505-a25fc32b45c8-kube-api-access-v28m2\") pod \"dns-default-2d6vb\" (UID: \"1b489c90-5e8b-4a49-8505-a25fc32b45c8\") " pod="openshift-dns/dns-default-2d6vb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109912 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a54581b-46f9-4453-b3c0-9cb55e3410f6-metrics-certs\") pod \"router-default-5444994796-mmrgn\" (UID: \"3a54581b-46f9-4453-b3c0-9cb55e3410f6\") " pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109952 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc57x\" (UniqueName: \"kubernetes.io/projected/d0d60e28-fafc-4490-802e-ee578094c9a3-kube-api-access-sc57x\") pod \"collect-profiles-29400075-dhldn\" (UID: \"d0d60e28-fafc-4490-802e-ee578094c9a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.109980 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/932ea52a-d515-4a68-b53a-84864d62fb37-serving-cert\") pod \"service-ca-operator-777779d784-b4lwp\" (UID: \"932ea52a-d515-4a68-b53a-84864d62fb37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.110001 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0d60e28-fafc-4490-802e-ee578094c9a3-config-volume\") pod \"collect-profiles-29400075-dhldn\" (UID: \"d0d60e28-fafc-4490-802e-ee578094c9a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.110046 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/94a29bbc-ed8f-46cf-9917-2ccba227dbcb-signing-key\") pod \"service-ca-9c57cc56f-5c5r4\" (UID: \"94a29bbc-ed8f-46cf-9917-2ccba227dbcb\") " pod="openshift-service-ca/service-ca-9c57cc56f-5c5r4" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.110097 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-etcd-client\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.110120 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0706d14f-f31b-4935-a1f5-6e1af9f11b13-trusted-ca\") pod \"ingress-operator-5b745b69d9-zcmdg\" (UID: \"0706d14f-f31b-4935-a1f5-6e1af9f11b13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.110515 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/22764299-441e-4ac3-986a-5a4c1a325816-images\") pod \"machine-config-operator-74547568cd-c4svp\" (UID: \"22764299-441e-4ac3-986a-5a4c1a325816\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.110601 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6faa23fa-42a4-499b-99b2-b02585cdf38f-cert\") pod \"ingress-canary-m57fw\" (UID: \"6faa23fa-42a4-499b-99b2-b02585cdf38f\") " pod="openshift-ingress-canary/ingress-canary-m57fw" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.110639 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/81133ff7-9de8-4958-b47c-7513df033b38-proxy-tls\") pod \"machine-config-controller-84d6567774-9dqfr\" (UID: \"81133ff7-9de8-4958-b47c-7513df033b38\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.110675 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29292c32-1a21-415c-83ce-1f526a0264b5-config\") pod \"kube-controller-manager-operator-78b949d7b-kgrtc\" (UID: \"29292c32-1a21-415c-83ce-1f526a0264b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.110736 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-config\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.110770 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c908f735-bb61-4a7c-a9be-4f354dfe332b-profile-collector-cert\") pod \"catalog-operator-68c6474976-lpbwg\" (UID: \"c908f735-bb61-4a7c-a9be-4f354dfe332b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.111671 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zbdt\" (UniqueName: \"kubernetes.io/projected/c908f735-bb61-4a7c-a9be-4f354dfe332b-kube-api-access-2zbdt\") pod \"catalog-operator-68c6474976-lpbwg\" (UID: \"c908f735-bb61-4a7c-a9be-4f354dfe332b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.112199 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.112393 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2v2lb\" (UID: \"94a5dac5-a893-481f-ab4e-bfbd4106b8c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.112476 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqbrm\" (UniqueName: \"kubernetes.io/projected/a9e3ed7b-d5c9-4d92-9c13-2ceba9741467-kube-api-access-kqbrm\") pod \"migrator-59844c95c7-nvgpx\" (UID: \"a9e3ed7b-d5c9-4d92-9c13-2ceba9741467\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nvgpx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.112511 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b489c90-5e8b-4a49-8505-a25fc32b45c8-config-volume\") pod \"dns-default-2d6vb\" (UID: \"1b489c90-5e8b-4a49-8505-a25fc32b45c8\") " pod="openshift-dns/dns-default-2d6vb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.112592 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/664e0a86-ebcc-4119-948c-feefad6325a6-socket-dir\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.112659 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jlml\" (UniqueName: \"kubernetes.io/projected/13575d67-6fe6-4ba5-991b-17ef3b2d3878-kube-api-access-9jlml\") pod \"packageserver-d55dfcdfc-gx962\" (UID: \"13575d67-6fe6-4ba5-991b-17ef3b2d3878\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.112716 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/664e0a86-ebcc-4119-948c-feefad6325a6-mountpoint-dir\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.112742 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/29292c32-1a21-415c-83ce-1f526a0264b5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kgrtc\" (UID: \"29292c32-1a21-415c-83ce-1f526a0264b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.112790 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/edb7cea2-f6eb-4043-91d9-c7abaef74455-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.112844 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/81133ff7-9de8-4958-b47c-7513df033b38-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9dqfr\" (UID: \"81133ff7-9de8-4958-b47c-7513df033b38\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.113259 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp8fx\" (UniqueName: \"kubernetes.io/projected/664e0a86-ebcc-4119-948c-feefad6325a6-kube-api-access-rp8fx\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.113301 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmrnp\" (UniqueName: \"kubernetes.io/projected/d594e8b8-8eb4-4cfa-b978-6fa66ca4c379-kube-api-access-fmrnp\") pod \"package-server-manager-789f6589d5-wsfw6\" (UID: \"d594e8b8-8eb4-4cfa-b978-6fa66ca4c379\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.113358 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/edb7cea2-f6eb-4043-91d9-c7abaef74455-trusted-ca\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.113418 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72380442-6dd0-45fd-82ce-07191fb9427e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-w2bpk\" (UID: \"72380442-6dd0-45fd-82ce-07191fb9427e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.113585 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932ea52a-d515-4a68-b53a-84864d62fb37-config\") pod \"service-ca-operator-777779d784-b4lwp\" (UID: \"932ea52a-d515-4a68-b53a-84864d62fb37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.113627 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/664e0a86-ebcc-4119-948c-feefad6325a6-plugins-dir\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.113683 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72380442-6dd0-45fd-82ce-07191fb9427e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-w2bpk\" (UID: \"72380442-6dd0-45fd-82ce-07191fb9427e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.113712 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98lb9\" (UniqueName: \"kubernetes.io/projected/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-kube-api-access-98lb9\") pod \"marketplace-operator-79b997595-2v2lb\" (UID: \"94a5dac5-a893-481f-ab4e-bfbd4106b8c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.113741 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/edb7cea2-f6eb-4043-91d9-c7abaef74455-registry-certificates\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.113794 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbf7x\" (UniqueName: \"kubernetes.io/projected/932ea52a-d515-4a68-b53a-84864d62fb37-kube-api-access-cbf7x\") pod \"service-ca-operator-777779d784-b4lwp\" (UID: \"932ea52a-d515-4a68-b53a-84864d62fb37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.113839 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c908f735-bb61-4a7c-a9be-4f354dfe332b-srv-cert\") pod \"catalog-operator-68c6474976-lpbwg\" (UID: \"c908f735-bb61-4a7c-a9be-4f354dfe332b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.114223 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/664e0a86-ebcc-4119-948c-feefad6325a6-csi-data-dir\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.114403 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.114918 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df575f42-a7d3-42d8-9d03-f5334bbbc35b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfd6f\" (UID: \"df575f42-a7d3-42d8-9d03-f5334bbbc35b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.115066 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb88d\" (UniqueName: \"kubernetes.io/projected/22764299-441e-4ac3-986a-5a4c1a325816-kube-api-access-cb88d\") pod \"machine-config-operator-74547568cd-c4svp\" (UID: \"22764299-441e-4ac3-986a-5a4c1a325816\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.115158 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-etcd-service-ca\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: E1124 17:29:10.115201 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:10.615177595 +0000 UTC m=+143.212845437 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.115252 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d594e8b8-8eb4-4cfa-b978-6fa66ca4c379-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wsfw6\" (UID: \"d594e8b8-8eb4-4cfa-b978-6fa66ca4c379\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.115302 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/24aea72e-3aea-48a5-ac6f-8ab8bb516b18-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tls4q\" (UID: \"24aea72e-3aea-48a5-ac6f-8ab8bb516b18\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tls4q" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.115391 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4c27c616-6ab4-40fb-857f-cfd4d507547e-srv-cert\") pod \"olm-operator-6b444d44fb-xlpkq\" (UID: \"4c27c616-6ab4-40fb-857f-cfd4d507547e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.115490 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/664e0a86-ebcc-4119-948c-feefad6325a6-registration-dir\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.115523 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4c27c616-6ab4-40fb-857f-cfd4d507547e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xlpkq\" (UID: \"4c27c616-6ab4-40fb-857f-cfd4d507547e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.115584 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1b489c90-5e8b-4a49-8505-a25fc32b45c8-metrics-tls\") pod \"dns-default-2d6vb\" (UID: \"1b489c90-5e8b-4a49-8505-a25fc32b45c8\") " pod="openshift-dns/dns-default-2d6vb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.115752 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vxr7\" (UniqueName: \"kubernetes.io/projected/3a54581b-46f9-4453-b3c0-9cb55e3410f6-kube-api-access-2vxr7\") pod \"router-default-5444994796-mmrgn\" (UID: \"3a54581b-46f9-4453-b3c0-9cb55e3410f6\") " pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.115765 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/edb7cea2-f6eb-4043-91d9-c7abaef74455-trusted-ca\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.115804 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3a54581b-46f9-4453-b3c0-9cb55e3410f6-default-certificate\") pod \"router-default-5444994796-mmrgn\" (UID: \"3a54581b-46f9-4453-b3c0-9cb55e3410f6\") " pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.115832 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5rjl\" (UniqueName: \"kubernetes.io/projected/94a29bbc-ed8f-46cf-9917-2ccba227dbcb-kube-api-access-l5rjl\") pod \"service-ca-9c57cc56f-5c5r4\" (UID: \"94a29bbc-ed8f-46cf-9917-2ccba227dbcb\") " pod="openshift-service-ca/service-ca-9c57cc56f-5c5r4" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.115859 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9478\" (UniqueName: \"kubernetes.io/projected/6faa23fa-42a4-499b-99b2-b02585cdf38f-kube-api-access-p9478\") pod \"ingress-canary-m57fw\" (UID: \"6faa23fa-42a4-499b-99b2-b02585cdf38f\") " pod="openshift-ingress-canary/ingress-canary-m57fw" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.115945 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3a54581b-46f9-4453-b3c0-9cb55e3410f6-stats-auth\") pod \"router-default-5444994796-mmrgn\" (UID: \"3a54581b-46f9-4453-b3c0-9cb55e3410f6\") " pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.116230 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-bound-sa-token\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.116286 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4319fe26-e7cc-4665-881a-6275268b6330-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ljn6r\" (UID: \"4319fe26-e7cc-4665-881a-6275268b6330\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ljn6r" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.116371 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/13575d67-6fe6-4ba5-991b-17ef3b2d3878-apiservice-cert\") pod \"packageserver-d55dfcdfc-gx962\" (UID: \"13575d67-6fe6-4ba5-991b-17ef3b2d3878\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.116417 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxnbk\" (UniqueName: \"kubernetes.io/projected/ab4ee277-dff6-4411-9fe1-72e56da6244b-kube-api-access-dxnbk\") pod \"machine-config-server-7sbnn\" (UID: \"ab4ee277-dff6-4411-9fe1-72e56da6244b\") " pod="openshift-machine-config-operator/machine-config-server-7sbnn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.116452 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnnlz\" (UniqueName: \"kubernetes.io/projected/24aea72e-3aea-48a5-ac6f-8ab8bb516b18-kube-api-access-rnnlz\") pod \"multus-admission-controller-857f4d67dd-tls4q\" (UID: \"24aea72e-3aea-48a5-ac6f-8ab8bb516b18\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tls4q" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.116529 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df575f42-a7d3-42d8-9d03-f5334bbbc35b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfd6f\" (UID: \"df575f42-a7d3-42d8-9d03-f5334bbbc35b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.117388 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/edb7cea2-f6eb-4043-91d9-c7abaef74455-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.117423 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/13575d67-6fe6-4ba5-991b-17ef3b2d3878-tmpfs\") pod \"packageserver-d55dfcdfc-gx962\" (UID: \"13575d67-6fe6-4ba5-991b-17ef3b2d3878\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.117460 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-etcd-ca\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.117515 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9b68\" (UniqueName: \"kubernetes.io/projected/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-kube-api-access-t9b68\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.117676 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/edb7cea2-f6eb-4043-91d9-c7abaef74455-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.117742 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/13575d67-6fe6-4ba5-991b-17ef3b2d3878-webhook-cert\") pod \"packageserver-d55dfcdfc-gx962\" (UID: \"13575d67-6fe6-4ba5-991b-17ef3b2d3878\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.118063 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/edb7cea2-f6eb-4043-91d9-c7abaef74455-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.118065 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/edb7cea2-f6eb-4043-91d9-c7abaef74455-registry-certificates\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.118157 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ab4ee277-dff6-4411-9fe1-72e56da6244b-certs\") pod \"machine-config-server-7sbnn\" (UID: \"ab4ee277-dff6-4411-9fe1-72e56da6244b\") " pod="openshift-machine-config-operator/machine-config-server-7sbnn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.118221 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-registry-tls\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.118295 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2v2lb\" (UID: \"94a5dac5-a893-481f-ab4e-bfbd4106b8c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.118340 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0d60e28-fafc-4490-802e-ee578094c9a3-secret-volume\") pod \"collect-profiles-29400075-dhldn\" (UID: \"d0d60e28-fafc-4490-802e-ee578094c9a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.118478 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22764299-441e-4ac3-986a-5a4c1a325816-auth-proxy-config\") pod \"machine-config-operator-74547568cd-c4svp\" (UID: \"22764299-441e-4ac3-986a-5a4c1a325816\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.125093 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-registry-tls\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.125966 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jmbl\" (UniqueName: \"kubernetes.io/projected/0706d14f-f31b-4935-a1f5-6e1af9f11b13-kube-api-access-7jmbl\") pod \"ingress-operator-5b745b69d9-zcmdg\" (UID: \"0706d14f-f31b-4935-a1f5-6e1af9f11b13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.126059 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/94a29bbc-ed8f-46cf-9917-2ccba227dbcb-signing-cabundle\") pod \"service-ca-9c57cc56f-5c5r4\" (UID: \"94a29bbc-ed8f-46cf-9917-2ccba227dbcb\") " pod="openshift-service-ca/service-ca-9c57cc56f-5c5r4" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.126134 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ab4ee277-dff6-4411-9fe1-72e56da6244b-node-bootstrap-token\") pod \"machine-config-server-7sbnn\" (UID: \"ab4ee277-dff6-4411-9fe1-72e56da6244b\") " pod="openshift-machine-config-operator/machine-config-server-7sbnn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.134438 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.150783 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.161409 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-qlpqd" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.176754 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9c6772a-c49e-441b-86a3-b76d1ea62823-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-chwdd\" (UID: \"e9c6772a-c49e-441b-86a3-b76d1ea62823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.211895 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqbq8\" (UniqueName: \"kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-kube-api-access-mqbq8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.226725 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:10 crc kubenswrapper[4808]: E1124 17:29:10.226963 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:10.726921033 +0000 UTC m=+143.324588845 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.227404 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc57x\" (UniqueName: \"kubernetes.io/projected/d0d60e28-fafc-4490-802e-ee578094c9a3-kube-api-access-sc57x\") pod \"collect-profiles-29400075-dhldn\" (UID: \"d0d60e28-fafc-4490-802e-ee578094c9a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.227506 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a54581b-46f9-4453-b3c0-9cb55e3410f6-metrics-certs\") pod \"router-default-5444994796-mmrgn\" (UID: \"3a54581b-46f9-4453-b3c0-9cb55e3410f6\") " pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.227786 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0d60e28-fafc-4490-802e-ee578094c9a3-config-volume\") pod \"collect-profiles-29400075-dhldn\" (UID: \"d0d60e28-fafc-4490-802e-ee578094c9a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.227870 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/932ea52a-d515-4a68-b53a-84864d62fb37-serving-cert\") pod \"service-ca-operator-777779d784-b4lwp\" (UID: \"932ea52a-d515-4a68-b53a-84864d62fb37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.227948 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/94a29bbc-ed8f-46cf-9917-2ccba227dbcb-signing-key\") pod \"service-ca-9c57cc56f-5c5r4\" (UID: \"94a29bbc-ed8f-46cf-9917-2ccba227dbcb\") " pod="openshift-service-ca/service-ca-9c57cc56f-5c5r4" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.228044 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-etcd-client\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.228131 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0706d14f-f31b-4935-a1f5-6e1af9f11b13-trusted-ca\") pod \"ingress-operator-5b745b69d9-zcmdg\" (UID: \"0706d14f-f31b-4935-a1f5-6e1af9f11b13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.228201 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/22764299-441e-4ac3-986a-5a4c1a325816-images\") pod \"machine-config-operator-74547568cd-c4svp\" (UID: \"22764299-441e-4ac3-986a-5a4c1a325816\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.228275 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/81133ff7-9de8-4958-b47c-7513df033b38-proxy-tls\") pod \"machine-config-controller-84d6567774-9dqfr\" (UID: \"81133ff7-9de8-4958-b47c-7513df033b38\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.228374 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29292c32-1a21-415c-83ce-1f526a0264b5-config\") pod \"kube-controller-manager-operator-78b949d7b-kgrtc\" (UID: \"29292c32-1a21-415c-83ce-1f526a0264b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.228452 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6faa23fa-42a4-499b-99b2-b02585cdf38f-cert\") pod \"ingress-canary-m57fw\" (UID: \"6faa23fa-42a4-499b-99b2-b02585cdf38f\") " pod="openshift-ingress-canary/ingress-canary-m57fw" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.228528 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-config\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.228606 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c908f735-bb61-4a7c-a9be-4f354dfe332b-profile-collector-cert\") pod \"catalog-operator-68c6474976-lpbwg\" (UID: \"c908f735-bb61-4a7c-a9be-4f354dfe332b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.228715 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zbdt\" (UniqueName: \"kubernetes.io/projected/c908f735-bb61-4a7c-a9be-4f354dfe332b-kube-api-access-2zbdt\") pod \"catalog-operator-68c6474976-lpbwg\" (UID: \"c908f735-bb61-4a7c-a9be-4f354dfe332b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.228859 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2v2lb\" (UID: \"94a5dac5-a893-481f-ab4e-bfbd4106b8c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.228963 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqbrm\" (UniqueName: \"kubernetes.io/projected/a9e3ed7b-d5c9-4d92-9c13-2ceba9741467-kube-api-access-kqbrm\") pod \"migrator-59844c95c7-nvgpx\" (UID: \"a9e3ed7b-d5c9-4d92-9c13-2ceba9741467\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nvgpx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.229082 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b489c90-5e8b-4a49-8505-a25fc32b45c8-config-volume\") pod \"dns-default-2d6vb\" (UID: \"1b489c90-5e8b-4a49-8505-a25fc32b45c8\") " pod="openshift-dns/dns-default-2d6vb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.229187 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/664e0a86-ebcc-4119-948c-feefad6325a6-socket-dir\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.229311 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jlml\" (UniqueName: \"kubernetes.io/projected/13575d67-6fe6-4ba5-991b-17ef3b2d3878-kube-api-access-9jlml\") pod \"packageserver-d55dfcdfc-gx962\" (UID: \"13575d67-6fe6-4ba5-991b-17ef3b2d3878\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.229416 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/664e0a86-ebcc-4119-948c-feefad6325a6-mountpoint-dir\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.229593 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/29292c32-1a21-415c-83ce-1f526a0264b5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kgrtc\" (UID: \"29292c32-1a21-415c-83ce-1f526a0264b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.229703 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/81133ff7-9de8-4958-b47c-7513df033b38-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9dqfr\" (UID: \"81133ff7-9de8-4958-b47c-7513df033b38\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.230006 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmrnp\" (UniqueName: \"kubernetes.io/projected/d594e8b8-8eb4-4cfa-b978-6fa66ca4c379-kube-api-access-fmrnp\") pod \"package-server-manager-789f6589d5-wsfw6\" (UID: \"d594e8b8-8eb4-4cfa-b978-6fa66ca4c379\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.230135 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp8fx\" (UniqueName: \"kubernetes.io/projected/664e0a86-ebcc-4119-948c-feefad6325a6-kube-api-access-rp8fx\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.230231 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72380442-6dd0-45fd-82ce-07191fb9427e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-w2bpk\" (UID: \"72380442-6dd0-45fd-82ce-07191fb9427e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.230330 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932ea52a-d515-4a68-b53a-84864d62fb37-config\") pod \"service-ca-operator-777779d784-b4lwp\" (UID: \"932ea52a-d515-4a68-b53a-84864d62fb37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.230425 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/664e0a86-ebcc-4119-948c-feefad6325a6-plugins-dir\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.230522 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72380442-6dd0-45fd-82ce-07191fb9427e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-w2bpk\" (UID: \"72380442-6dd0-45fd-82ce-07191fb9427e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.230771 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98lb9\" (UniqueName: \"kubernetes.io/projected/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-kube-api-access-98lb9\") pod \"marketplace-operator-79b997595-2v2lb\" (UID: \"94a5dac5-a893-481f-ab4e-bfbd4106b8c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.230891 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbf7x\" (UniqueName: \"kubernetes.io/projected/932ea52a-d515-4a68-b53a-84864d62fb37-kube-api-access-cbf7x\") pod \"service-ca-operator-777779d784-b4lwp\" (UID: \"932ea52a-d515-4a68-b53a-84864d62fb37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.230995 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c908f735-bb61-4a7c-a9be-4f354dfe332b-srv-cert\") pod \"catalog-operator-68c6474976-lpbwg\" (UID: \"c908f735-bb61-4a7c-a9be-4f354dfe332b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.231106 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/664e0a86-ebcc-4119-948c-feefad6325a6-csi-data-dir\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.231232 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.231343 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df575f42-a7d3-42d8-9d03-f5334bbbc35b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfd6f\" (UID: \"df575f42-a7d3-42d8-9d03-f5334bbbc35b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.231449 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-etcd-service-ca\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.231543 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d594e8b8-8eb4-4cfa-b978-6fa66ca4c379-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wsfw6\" (UID: \"d594e8b8-8eb4-4cfa-b978-6fa66ca4c379\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.231650 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/24aea72e-3aea-48a5-ac6f-8ab8bb516b18-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tls4q\" (UID: \"24aea72e-3aea-48a5-ac6f-8ab8bb516b18\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tls4q" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.231760 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4c27c616-6ab4-40fb-857f-cfd4d507547e-srv-cert\") pod \"olm-operator-6b444d44fb-xlpkq\" (UID: \"4c27c616-6ab4-40fb-857f-cfd4d507547e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.231848 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb88d\" (UniqueName: \"kubernetes.io/projected/22764299-441e-4ac3-986a-5a4c1a325816-kube-api-access-cb88d\") pod \"machine-config-operator-74547568cd-c4svp\" (UID: \"22764299-441e-4ac3-986a-5a4c1a325816\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.231927 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/664e0a86-ebcc-4119-948c-feefad6325a6-registration-dir\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.231982 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29292c32-1a21-415c-83ce-1f526a0264b5-config\") pod \"kube-controller-manager-operator-78b949d7b-kgrtc\" (UID: \"29292c32-1a21-415c-83ce-1f526a0264b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.232399 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0706d14f-f31b-4935-a1f5-6e1af9f11b13-trusted-ca\") pod \"ingress-operator-5b745b69d9-zcmdg\" (UID: \"0706d14f-f31b-4935-a1f5-6e1af9f11b13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.232631 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/664e0a86-ebcc-4119-948c-feefad6325a6-csi-data-dir\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.229137 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0d60e28-fafc-4490-802e-ee578094c9a3-config-volume\") pod \"collect-profiles-29400075-dhldn\" (UID: \"d0d60e28-fafc-4490-802e-ee578094c9a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" Nov 24 17:29:10 crc kubenswrapper[4808]: E1124 17:29:10.233097 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:10.733078336 +0000 UTC m=+143.330746158 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.233722 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2v2lb\" (UID: \"94a5dac5-a893-481f-ab4e-bfbd4106b8c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.231445 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/932ea52a-d515-4a68-b53a-84864d62fb37-serving-cert\") pod \"service-ca-operator-777779d784-b4lwp\" (UID: \"932ea52a-d515-4a68-b53a-84864d62fb37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.233811 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a54581b-46f9-4453-b3c0-9cb55e3410f6-metrics-certs\") pod \"router-default-5444994796-mmrgn\" (UID: \"3a54581b-46f9-4453-b3c0-9cb55e3410f6\") " pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.231260 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/664e0a86-ebcc-4119-948c-feefad6325a6-socket-dir\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.229429 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-bound-sa-token\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.233998 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/664e0a86-ebcc-4119-948c-feefad6325a6-mountpoint-dir\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.229455 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/22764299-441e-4ac3-986a-5a4c1a325816-images\") pod \"machine-config-operator-74547568cd-c4svp\" (UID: \"22764299-441e-4ac3-986a-5a4c1a325816\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.234144 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/664e0a86-ebcc-4119-948c-feefad6325a6-registration-dir\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.234709 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-etcd-service-ca\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.234891 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72380442-6dd0-45fd-82ce-07191fb9427e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-w2bpk\" (UID: \"72380442-6dd0-45fd-82ce-07191fb9427e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.235013 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c908f735-bb61-4a7c-a9be-4f354dfe332b-profile-collector-cert\") pod \"catalog-operator-68c6474976-lpbwg\" (UID: \"c908f735-bb61-4a7c-a9be-4f354dfe332b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.235130 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/664e0a86-ebcc-4119-948c-feefad6325a6-plugins-dir\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.235514 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932ea52a-d515-4a68-b53a-84864d62fb37-config\") pod \"service-ca-operator-777779d784-b4lwp\" (UID: \"932ea52a-d515-4a68-b53a-84864d62fb37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.235700 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-config\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.236657 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b489c90-5e8b-4a49-8505-a25fc32b45c8-config-volume\") pod \"dns-default-2d6vb\" (UID: \"1b489c90-5e8b-4a49-8505-a25fc32b45c8\") " pod="openshift-dns/dns-default-2d6vb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.236835 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4c27c616-6ab4-40fb-857f-cfd4d507547e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xlpkq\" (UID: \"4c27c616-6ab4-40fb-857f-cfd4d507547e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.237434 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1b489c90-5e8b-4a49-8505-a25fc32b45c8-metrics-tls\") pod \"dns-default-2d6vb\" (UID: \"1b489c90-5e8b-4a49-8505-a25fc32b45c8\") " pod="openshift-dns/dns-default-2d6vb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.237557 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vxr7\" (UniqueName: \"kubernetes.io/projected/3a54581b-46f9-4453-b3c0-9cb55e3410f6-kube-api-access-2vxr7\") pod \"router-default-5444994796-mmrgn\" (UID: \"3a54581b-46f9-4453-b3c0-9cb55e3410f6\") " pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.237719 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5rjl\" (UniqueName: \"kubernetes.io/projected/94a29bbc-ed8f-46cf-9917-2ccba227dbcb-kube-api-access-l5rjl\") pod \"service-ca-9c57cc56f-5c5r4\" (UID: \"94a29bbc-ed8f-46cf-9917-2ccba227dbcb\") " pod="openshift-service-ca/service-ca-9c57cc56f-5c5r4" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.237826 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9478\" (UniqueName: \"kubernetes.io/projected/6faa23fa-42a4-499b-99b2-b02585cdf38f-kube-api-access-p9478\") pod \"ingress-canary-m57fw\" (UID: \"6faa23fa-42a4-499b-99b2-b02585cdf38f\") " pod="openshift-ingress-canary/ingress-canary-m57fw" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.237954 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3a54581b-46f9-4453-b3c0-9cb55e3410f6-default-certificate\") pod \"router-default-5444994796-mmrgn\" (UID: \"3a54581b-46f9-4453-b3c0-9cb55e3410f6\") " pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.238071 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3a54581b-46f9-4453-b3c0-9cb55e3410f6-stats-auth\") pod \"router-default-5444994796-mmrgn\" (UID: \"3a54581b-46f9-4453-b3c0-9cb55e3410f6\") " pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.238185 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4319fe26-e7cc-4665-881a-6275268b6330-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ljn6r\" (UID: \"4319fe26-e7cc-4665-881a-6275268b6330\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ljn6r" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.238301 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/13575d67-6fe6-4ba5-991b-17ef3b2d3878-apiservice-cert\") pod \"packageserver-d55dfcdfc-gx962\" (UID: \"13575d67-6fe6-4ba5-991b-17ef3b2d3878\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.238409 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxnbk\" (UniqueName: \"kubernetes.io/projected/ab4ee277-dff6-4411-9fe1-72e56da6244b-kube-api-access-dxnbk\") pod \"machine-config-server-7sbnn\" (UID: \"ab4ee277-dff6-4411-9fe1-72e56da6244b\") " pod="openshift-machine-config-operator/machine-config-server-7sbnn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.238536 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnnlz\" (UniqueName: \"kubernetes.io/projected/24aea72e-3aea-48a5-ac6f-8ab8bb516b18-kube-api-access-rnnlz\") pod \"multus-admission-controller-857f4d67dd-tls4q\" (UID: \"24aea72e-3aea-48a5-ac6f-8ab8bb516b18\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tls4q" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.238664 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df575f42-a7d3-42d8-9d03-f5334bbbc35b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfd6f\" (UID: \"df575f42-a7d3-42d8-9d03-f5334bbbc35b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.238778 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/13575d67-6fe6-4ba5-991b-17ef3b2d3878-tmpfs\") pod \"packageserver-d55dfcdfc-gx962\" (UID: \"13575d67-6fe6-4ba5-991b-17ef3b2d3878\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.238878 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-etcd-ca\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.238987 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9b68\" (UniqueName: \"kubernetes.io/projected/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-kube-api-access-t9b68\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.239149 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/13575d67-6fe6-4ba5-991b-17ef3b2d3878-webhook-cert\") pod \"packageserver-d55dfcdfc-gx962\" (UID: \"13575d67-6fe6-4ba5-991b-17ef3b2d3878\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.239244 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ab4ee277-dff6-4411-9fe1-72e56da6244b-certs\") pod \"machine-config-server-7sbnn\" (UID: \"ab4ee277-dff6-4411-9fe1-72e56da6244b\") " pod="openshift-machine-config-operator/machine-config-server-7sbnn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.239345 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0d60e28-fafc-4490-802e-ee578094c9a3-secret-volume\") pod \"collect-profiles-29400075-dhldn\" (UID: \"d0d60e28-fafc-4490-802e-ee578094c9a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.239462 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2v2lb\" (UID: \"94a5dac5-a893-481f-ab4e-bfbd4106b8c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.239774 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22764299-441e-4ac3-986a-5a4c1a325816-auth-proxy-config\") pod \"machine-config-operator-74547568cd-c4svp\" (UID: \"22764299-441e-4ac3-986a-5a4c1a325816\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.241252 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jmbl\" (UniqueName: \"kubernetes.io/projected/0706d14f-f31b-4935-a1f5-6e1af9f11b13-kube-api-access-7jmbl\") pod \"ingress-operator-5b745b69d9-zcmdg\" (UID: \"0706d14f-f31b-4935-a1f5-6e1af9f11b13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.242784 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/94a29bbc-ed8f-46cf-9917-2ccba227dbcb-signing-cabundle\") pod \"service-ca-9c57cc56f-5c5r4\" (UID: \"94a29bbc-ed8f-46cf-9917-2ccba227dbcb\") " pod="openshift-service-ca/service-ca-9c57cc56f-5c5r4" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.242828 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ab4ee277-dff6-4411-9fe1-72e56da6244b-node-bootstrap-token\") pod \"machine-config-server-7sbnn\" (UID: \"ab4ee277-dff6-4411-9fe1-72e56da6244b\") " pod="openshift-machine-config-operator/machine-config-server-7sbnn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.242859 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/22764299-441e-4ac3-986a-5a4c1a325816-proxy-tls\") pod \"machine-config-operator-74547568cd-c4svp\" (UID: \"22764299-441e-4ac3-986a-5a4c1a325816\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.242885 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7plk6\" (UniqueName: \"kubernetes.io/projected/4319fe26-e7cc-4665-881a-6275268b6330-kube-api-access-7plk6\") pod \"control-plane-machine-set-operator-78cbb6b69f-ljn6r\" (UID: \"4319fe26-e7cc-4665-881a-6275268b6330\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ljn6r" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.242915 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df575f42-a7d3-42d8-9d03-f5334bbbc35b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfd6f\" (UID: \"df575f42-a7d3-42d8-9d03-f5334bbbc35b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.242940 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29292c32-1a21-415c-83ce-1f526a0264b5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kgrtc\" (UID: \"29292c32-1a21-415c-83ce-1f526a0264b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.242974 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a54581b-46f9-4453-b3c0-9cb55e3410f6-service-ca-bundle\") pod \"router-default-5444994796-mmrgn\" (UID: \"3a54581b-46f9-4453-b3c0-9cb55e3410f6\") " pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.242998 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68fp2\" (UniqueName: \"kubernetes.io/projected/4c27c616-6ab4-40fb-857f-cfd4d507547e-kube-api-access-68fp2\") pod \"olm-operator-6b444d44fb-xlpkq\" (UID: \"4c27c616-6ab4-40fb-857f-cfd4d507547e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.243053 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0706d14f-f31b-4935-a1f5-6e1af9f11b13-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zcmdg\" (UID: \"0706d14f-f31b-4935-a1f5-6e1af9f11b13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.243080 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m75ks\" (UniqueName: \"kubernetes.io/projected/81133ff7-9de8-4958-b47c-7513df033b38-kube-api-access-m75ks\") pod \"machine-config-controller-84d6567774-9dqfr\" (UID: \"81133ff7-9de8-4958-b47c-7513df033b38\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.243109 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlzm7\" (UniqueName: \"kubernetes.io/projected/72380442-6dd0-45fd-82ce-07191fb9427e-kube-api-access-nlzm7\") pod \"kube-storage-version-migrator-operator-b67b599dd-w2bpk\" (UID: \"72380442-6dd0-45fd-82ce-07191fb9427e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.243141 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-serving-cert\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.243173 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0706d14f-f31b-4935-a1f5-6e1af9f11b13-metrics-tls\") pod \"ingress-operator-5b745b69d9-zcmdg\" (UID: \"0706d14f-f31b-4935-a1f5-6e1af9f11b13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.243201 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v28m2\" (UniqueName: \"kubernetes.io/projected/1b489c90-5e8b-4a49-8505-a25fc32b45c8-kube-api-access-v28m2\") pod \"dns-default-2d6vb\" (UID: \"1b489c90-5e8b-4a49-8505-a25fc32b45c8\") " pod="openshift-dns/dns-default-2d6vb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.238785 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/81133ff7-9de8-4958-b47c-7513df033b38-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9dqfr\" (UID: \"81133ff7-9de8-4958-b47c-7513df033b38\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.242656 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22764299-441e-4ac3-986a-5a4c1a325816-auth-proxy-config\") pod \"machine-config-operator-74547568cd-c4svp\" (UID: \"22764299-441e-4ac3-986a-5a4c1a325816\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.244182 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df575f42-a7d3-42d8-9d03-f5334bbbc35b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfd6f\" (UID: \"df575f42-a7d3-42d8-9d03-f5334bbbc35b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.244184 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/94a29bbc-ed8f-46cf-9917-2ccba227dbcb-signing-cabundle\") pod \"service-ca-9c57cc56f-5c5r4\" (UID: \"94a29bbc-ed8f-46cf-9917-2ccba227dbcb\") " pod="openshift-service-ca/service-ca-9c57cc56f-5c5r4" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.241205 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-etcd-ca\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.240595 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/13575d67-6fe6-4ba5-991b-17ef3b2d3878-tmpfs\") pod \"packageserver-d55dfcdfc-gx962\" (UID: \"13575d67-6fe6-4ba5-991b-17ef3b2d3878\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.245457 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72380442-6dd0-45fd-82ce-07191fb9427e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-w2bpk\" (UID: \"72380442-6dd0-45fd-82ce-07191fb9427e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.245969 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4c27c616-6ab4-40fb-857f-cfd4d507547e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xlpkq\" (UID: \"4c27c616-6ab4-40fb-857f-cfd4d507547e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.246889 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a54581b-46f9-4453-b3c0-9cb55e3410f6-service-ca-bundle\") pod \"router-default-5444994796-mmrgn\" (UID: \"3a54581b-46f9-4453-b3c0-9cb55e3410f6\") " pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.249636 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/13575d67-6fe6-4ba5-991b-17ef3b2d3878-webhook-cert\") pod \"packageserver-d55dfcdfc-gx962\" (UID: \"13575d67-6fe6-4ba5-991b-17ef3b2d3878\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.249775 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d594e8b8-8eb4-4cfa-b978-6fa66ca4c379-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wsfw6\" (UID: \"d594e8b8-8eb4-4cfa-b978-6fa66ca4c379\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.250326 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/13575d67-6fe6-4ba5-991b-17ef3b2d3878-apiservice-cert\") pod \"packageserver-d55dfcdfc-gx962\" (UID: \"13575d67-6fe6-4ba5-991b-17ef3b2d3878\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.251143 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/81133ff7-9de8-4958-b47c-7513df033b38-proxy-tls\") pod \"machine-config-controller-84d6567774-9dqfr\" (UID: \"81133ff7-9de8-4958-b47c-7513df033b38\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.251359 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-serving-cert\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.251600 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1b489c90-5e8b-4a49-8505-a25fc32b45c8-metrics-tls\") pod \"dns-default-2d6vb\" (UID: \"1b489c90-5e8b-4a49-8505-a25fc32b45c8\") " pod="openshift-dns/dns-default-2d6vb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.251766 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6faa23fa-42a4-499b-99b2-b02585cdf38f-cert\") pod \"ingress-canary-m57fw\" (UID: \"6faa23fa-42a4-499b-99b2-b02585cdf38f\") " pod="openshift-ingress-canary/ingress-canary-m57fw" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.252091 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ab4ee277-dff6-4411-9fe1-72e56da6244b-node-bootstrap-token\") pod \"machine-config-server-7sbnn\" (UID: \"ab4ee277-dff6-4411-9fe1-72e56da6244b\") " pod="openshift-machine-config-operator/machine-config-server-7sbnn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.252472 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0d60e28-fafc-4490-802e-ee578094c9a3-secret-volume\") pod \"collect-profiles-29400075-dhldn\" (UID: \"d0d60e28-fafc-4490-802e-ee578094c9a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.252846 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-etcd-client\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.252866 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3a54581b-46f9-4453-b3c0-9cb55e3410f6-stats-auth\") pod \"router-default-5444994796-mmrgn\" (UID: \"3a54581b-46f9-4453-b3c0-9cb55e3410f6\") " pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.253323 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/24aea72e-3aea-48a5-ac6f-8ab8bb516b18-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tls4q\" (UID: \"24aea72e-3aea-48a5-ac6f-8ab8bb516b18\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tls4q" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.253323 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df575f42-a7d3-42d8-9d03-f5334bbbc35b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfd6f\" (UID: \"df575f42-a7d3-42d8-9d03-f5334bbbc35b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.257221 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4c27c616-6ab4-40fb-857f-cfd4d507547e-srv-cert\") pod \"olm-operator-6b444d44fb-xlpkq\" (UID: \"4c27c616-6ab4-40fb-857f-cfd4d507547e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.257957 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c908f735-bb61-4a7c-a9be-4f354dfe332b-srv-cert\") pod \"catalog-operator-68c6474976-lpbwg\" (UID: \"c908f735-bb61-4a7c-a9be-4f354dfe332b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.261584 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ab4ee277-dff6-4411-9fe1-72e56da6244b-certs\") pod \"machine-config-server-7sbnn\" (UID: \"ab4ee277-dff6-4411-9fe1-72e56da6244b\") " pod="openshift-machine-config-operator/machine-config-server-7sbnn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.261817 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2v2lb\" (UID: \"94a5dac5-a893-481f-ab4e-bfbd4106b8c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.262836 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0706d14f-f31b-4935-a1f5-6e1af9f11b13-metrics-tls\") pod \"ingress-operator-5b745b69d9-zcmdg\" (UID: \"0706d14f-f31b-4935-a1f5-6e1af9f11b13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.265318 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4319fe26-e7cc-4665-881a-6275268b6330-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ljn6r\" (UID: \"4319fe26-e7cc-4665-881a-6275268b6330\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ljn6r" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.270535 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/22764299-441e-4ac3-986a-5a4c1a325816-proxy-tls\") pod \"machine-config-operator-74547568cd-c4svp\" (UID: \"22764299-441e-4ac3-986a-5a4c1a325816\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.270932 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/94a29bbc-ed8f-46cf-9917-2ccba227dbcb-signing-key\") pod \"service-ca-9c57cc56f-5c5r4\" (UID: \"94a29bbc-ed8f-46cf-9917-2ccba227dbcb\") " pod="openshift-service-ca/service-ca-9c57cc56f-5c5r4" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.274411 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29292c32-1a21-415c-83ce-1f526a0264b5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kgrtc\" (UID: \"29292c32-1a21-415c-83ce-1f526a0264b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.301317 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3a54581b-46f9-4453-b3c0-9cb55e3410f6-default-certificate\") pod \"router-default-5444994796-mmrgn\" (UID: \"3a54581b-46f9-4453-b3c0-9cb55e3410f6\") " pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.312846 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zbdt\" (UniqueName: \"kubernetes.io/projected/c908f735-bb61-4a7c-a9be-4f354dfe332b-kube-api-access-2zbdt\") pod \"catalog-operator-68c6474976-lpbwg\" (UID: \"c908f735-bb61-4a7c-a9be-4f354dfe332b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.318092 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc57x\" (UniqueName: \"kubernetes.io/projected/d0d60e28-fafc-4490-802e-ee578094c9a3-kube-api-access-sc57x\") pod \"collect-profiles-29400075-dhldn\" (UID: \"d0d60e28-fafc-4490-802e-ee578094c9a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.329862 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.336529 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98lb9\" (UniqueName: \"kubernetes.io/projected/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-kube-api-access-98lb9\") pod \"marketplace-operator-79b997595-2v2lb\" (UID: \"94a5dac5-a893-481f-ab4e-bfbd4106b8c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.339386 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df575f42-a7d3-42d8-9d03-f5334bbbc35b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfd6f\" (UID: \"df575f42-a7d3-42d8-9d03-f5334bbbc35b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.344354 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:10 crc kubenswrapper[4808]: E1124 17:29:10.345065 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:10.845003789 +0000 UTC m=+143.442671581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.358603 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbf7x\" (UniqueName: \"kubernetes.io/projected/932ea52a-d515-4a68-b53a-84864d62fb37-kube-api-access-cbf7x\") pod \"service-ca-operator-777779d784-b4lwp\" (UID: \"932ea52a-d515-4a68-b53a-84864d62fb37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.377251 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqbrm\" (UniqueName: \"kubernetes.io/projected/a9e3ed7b-d5c9-4d92-9c13-2ceba9741467-kube-api-access-kqbrm\") pod \"migrator-59844c95c7-nvgpx\" (UID: \"a9e3ed7b-d5c9-4d92-9c13-2ceba9741467\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nvgpx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.404489 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jlml\" (UniqueName: \"kubernetes.io/projected/13575d67-6fe6-4ba5-991b-17ef3b2d3878-kube-api-access-9jlml\") pod \"packageserver-d55dfcdfc-gx962\" (UID: \"13575d67-6fe6-4ba5-991b-17ef3b2d3878\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.419509 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/29292c32-1a21-415c-83ce-1f526a0264b5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kgrtc\" (UID: \"29292c32-1a21-415c-83ce-1f526a0264b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.429444 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb88d\" (UniqueName: \"kubernetes.io/projected/22764299-441e-4ac3-986a-5a4c1a325816-kube-api-access-cb88d\") pod \"machine-config-operator-74547568cd-c4svp\" (UID: \"22764299-441e-4ac3-986a-5a4c1a325816\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.447674 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: E1124 17:29:10.448352 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:10.948330106 +0000 UTC m=+143.545997908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.457801 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmrnp\" (UniqueName: \"kubernetes.io/projected/d594e8b8-8eb4-4cfa-b978-6fa66ca4c379-kube-api-access-fmrnp\") pod \"package-server-manager-789f6589d5-wsfw6\" (UID: \"d594e8b8-8eb4-4cfa-b978-6fa66ca4c379\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.471315 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.486293 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.492672 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp8fx\" (UniqueName: \"kubernetes.io/projected/664e0a86-ebcc-4119-948c-feefad6325a6-kube-api-access-rp8fx\") pod \"csi-hostpathplugin-wvj5k\" (UID: \"664e0a86-ebcc-4119-948c-feefad6325a6\") " pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.498696 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vxr7\" (UniqueName: \"kubernetes.io/projected/3a54581b-46f9-4453-b3c0-9cb55e3410f6-kube-api-access-2vxr7\") pod \"router-default-5444994796-mmrgn\" (UID: \"3a54581b-46f9-4453-b3c0-9cb55e3410f6\") " pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.499359 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.517553 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.523955 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5rjl\" (UniqueName: \"kubernetes.io/projected/94a29bbc-ed8f-46cf-9917-2ccba227dbcb-kube-api-access-l5rjl\") pod \"service-ca-9c57cc56f-5c5r4\" (UID: \"94a29bbc-ed8f-46cf-9917-2ccba227dbcb\") " pod="openshift-service-ca/service-ca-9c57cc56f-5c5r4" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.532515 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.550745 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:10 crc kubenswrapper[4808]: E1124 17:29:10.551390 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:11.051369484 +0000 UTC m=+143.649037286 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.551692 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5c5r4" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.552642 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk"] Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.554939 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9478\" (UniqueName: \"kubernetes.io/projected/6faa23fa-42a4-499b-99b2-b02585cdf38f-kube-api-access-p9478\") pod \"ingress-canary-m57fw\" (UID: \"6faa23fa-42a4-499b-99b2-b02585cdf38f\") " pod="openshift-ingress-canary/ingress-canary-m57fw" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.560164 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxnbk\" (UniqueName: \"kubernetes.io/projected/ab4ee277-dff6-4411-9fe1-72e56da6244b-kube-api-access-dxnbk\") pod \"machine-config-server-7sbnn\" (UID: \"ab4ee277-dff6-4411-9fe1-72e56da6244b\") " pod="openshift-machine-config-operator/machine-config-server-7sbnn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.575824 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.576964 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9b68\" (UniqueName: \"kubernetes.io/projected/42e9945a-6923-4ada-b2f3-8aaaa3f853fd-kube-api-access-t9b68\") pod \"etcd-operator-b45778765-vkxbx\" (UID: \"42e9945a-6923-4ada-b2f3-8aaaa3f853fd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.585827 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" event={"ID":"31954b89-9642-4b55-b21e-79882b30b726","Type":"ContainerStarted","Data":"f37f56bdf2b293663f28edac5f500a763a42c2a2a0edae8a045345ddd2704f6b"} Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.587037 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.587330 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.600988 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnnlz\" (UniqueName: \"kubernetes.io/projected/24aea72e-3aea-48a5-ac6f-8ab8bb516b18-kube-api-access-rnnlz\") pod \"multus-admission-controller-857f4d67dd-tls4q\" (UID: \"24aea72e-3aea-48a5-ac6f-8ab8bb516b18\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tls4q" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.601010 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.606852 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nvgpx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.608265 4808 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-vwchc container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.608362 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" podUID="31954b89-9642-4b55-b21e-79882b30b726" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.614203 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" event={"ID":"12cfc041-5c29-43cb-9087-c1cbfb333f8b","Type":"ContainerStarted","Data":"cd5f452d5f37af215eb39b2fe3549cf1ae3ca346d228683e2188d0d69f41c4ee"} Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.619153 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v28m2\" (UniqueName: \"kubernetes.io/projected/1b489c90-5e8b-4a49-8505-a25fc32b45c8-kube-api-access-v28m2\") pod \"dns-default-2d6vb\" (UID: \"1b489c90-5e8b-4a49-8505-a25fc32b45c8\") " pod="openshift-dns/dns-default-2d6vb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.620453 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" event={"ID":"d54d24ca-c6c4-4313-8f22-908ce8939026","Type":"ContainerStarted","Data":"d1454e90d842bf20ccb69886a72cd856c58bad16e0b96db463c624323c3def86"} Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.621796 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.638756 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7plk6\" (UniqueName: \"kubernetes.io/projected/4319fe26-e7cc-4665-881a-6275268b6330-kube-api-access-7plk6\") pod \"control-plane-machine-set-operator-78cbb6b69f-ljn6r\" (UID: \"4319fe26-e7cc-4665-881a-6275268b6330\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ljn6r" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.641900 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.653754 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: E1124 17:29:10.654743 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:11.154722692 +0000 UTC m=+143.752390484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.658579 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-tls4q" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.679126 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-2d6vb" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.679659 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68fp2\" (UniqueName: \"kubernetes.io/projected/4c27c616-6ab4-40fb-857f-cfd4d507547e-kube-api-access-68fp2\") pod \"olm-operator-6b444d44fb-xlpkq\" (UID: \"4c27c616-6ab4-40fb-857f-cfd4d507547e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.688672 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m75ks\" (UniqueName: \"kubernetes.io/projected/81133ff7-9de8-4958-b47c-7513df033b38-kube-api-access-m75ks\") pod \"machine-config-controller-84d6567774-9dqfr\" (UID: \"81133ff7-9de8-4958-b47c-7513df033b38\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.691160 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" event={"ID":"a2ec940a-90eb-48d1-82fc-f17260545e94","Type":"ContainerStarted","Data":"f3c421982761f6f681348bb9aba79fe38f2a0182f659defba3a099ba7585a976"} Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.691540 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.705894 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.723385 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7sbnn" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.723703 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-m57fw" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.723757 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0706d14f-f31b-4935-a1f5-6e1af9f11b13-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zcmdg\" (UID: \"0706d14f-f31b-4935-a1f5-6e1af9f11b13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.727639 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-brt82" event={"ID":"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d","Type":"ContainerStarted","Data":"380916c934587ca527f21d29d3e3721f2b54c362661e7e73e5643f4806db8801"} Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.731811 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlzm7\" (UniqueName: \"kubernetes.io/projected/72380442-6dd0-45fd-82ce-07191fb9427e-kube-api-access-nlzm7\") pod \"kube-storage-version-migrator-operator-b67b599dd-w2bpk\" (UID: \"72380442-6dd0-45fd-82ce-07191fb9427e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.736589 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv" event={"ID":"1edf7884-088e-497d-a952-76b36a7d867e","Type":"ContainerStarted","Data":"91194275fe9bc4176d0192570fdf6f340267b63f547ca8014c3d28cd1565475f"} Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.742231 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dztgx" event={"ID":"36b5e49a-573b-41ca-aa12-5c991b910a72","Type":"ContainerStarted","Data":"edf1f671d94f178629689170c57e9ebe8ecb5b268bba0169191e93c47cc82e98"} Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.742295 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dztgx" event={"ID":"36b5e49a-573b-41ca-aa12-5c991b910a72","Type":"ContainerStarted","Data":"cff6156addb9d9119726890f9df2595d245f205398daf2aab92ec8ffa9ea6a97"} Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.755512 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:10 crc kubenswrapper[4808]: E1124 17:29:10.755946 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:11.255914105 +0000 UTC m=+143.853581907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.756106 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: E1124 17:29:10.756727 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:11.256703639 +0000 UTC m=+143.854371441 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.763194 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jmbl\" (UniqueName: \"kubernetes.io/projected/0706d14f-f31b-4935-a1f5-6e1af9f11b13-kube-api-access-7jmbl\") pod \"ingress-operator-5b745b69d9-zcmdg\" (UID: \"0706d14f-f31b-4935-a1f5-6e1af9f11b13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.770739 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" event={"ID":"8b35999c-bfc9-44bd-bca5-a97e645a7ed9","Type":"ContainerStarted","Data":"5622cd8e6f418847d521b09ea2fd29f0d5b655eae07fbc09d3177fb71a09cc69"} Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.770796 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" event={"ID":"8b35999c-bfc9-44bd-bca5-a97e645a7ed9","Type":"ContainerStarted","Data":"9c7391c6afafc4ce7ba2a4c0b6f63f12a9b7f674b1f167c92b0b238f1721a13f"} Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.778083 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn"] Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.780541 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-qlpqd"] Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.778599 4808 patch_prober.go:28] interesting pod/downloads-7954f5f757-6xnwl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.781531 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6xnwl" podUID="47734a14-98f7-4be3-89a3-74f5a13ab3e3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.780515 4808 patch_prober.go:28] interesting pod/console-operator-58897d9998-z2qxd container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.781635 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-z2qxd" podUID="aac68768-0e54-4b67-a919-eaa55380171f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.792161 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.793146 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.806498 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.824858 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.842417 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.860992 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:10 crc kubenswrapper[4808]: E1124 17:29:10.861472 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:11.361450668 +0000 UTC m=+143.959118470 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.863369 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:10 crc kubenswrapper[4808]: E1124 17:29:10.866268 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:11.366246761 +0000 UTC m=+143.963914563 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.870839 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ljn6r" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.913462 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr" Nov 24 17:29:10 crc kubenswrapper[4808]: I1124 17:29:10.971496 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:10 crc kubenswrapper[4808]: E1124 17:29:10.971957 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:11.471939218 +0000 UTC m=+144.069607020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.051028 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg"] Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.073321 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:11 crc kubenswrapper[4808]: E1124 17:29:11.073695 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:11.573682718 +0000 UTC m=+144.171350520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.116275 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.153152 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f"] Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.162275 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd"] Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.174313 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:11 crc kubenswrapper[4808]: E1124 17:29:11.174679 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:11.674661195 +0000 UTC m=+144.272328997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.217752 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962"] Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.282512 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:11 crc kubenswrapper[4808]: E1124 17:29:11.282958 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:11.782933969 +0000 UTC m=+144.380601781 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.381206 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" podStartSLOduration=123.381180735 podStartE2EDuration="2m3.381180735s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:11.324070744 +0000 UTC m=+143.921738546" watchObservedRunningTime="2025-11-24 17:29:11.381180735 +0000 UTC m=+143.978848547" Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.388954 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:11 crc kubenswrapper[4808]: E1124 17:29:11.389435 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:11.88941752 +0000 UTC m=+144.487085322 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.499042 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:11 crc kubenswrapper[4808]: E1124 17:29:11.499595 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:11.99956395 +0000 UTC m=+144.597231752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.559477 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" podStartSLOduration=123.559441743 podStartE2EDuration="2m3.559441743s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:11.557885237 +0000 UTC m=+144.155553039" watchObservedRunningTime="2025-11-24 17:29:11.559441743 +0000 UTC m=+144.157109545" Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.602350 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:11 crc kubenswrapper[4808]: E1124 17:29:11.602863 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:12.102839706 +0000 UTC m=+144.700507508 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.645326 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dztgx" podStartSLOduration=123.64529901 podStartE2EDuration="2m3.64529901s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:11.603203356 +0000 UTC m=+144.200871168" watchObservedRunningTime="2025-11-24 17:29:11.64529901 +0000 UTC m=+144.242966822" Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.684812 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-qrcr6" podStartSLOduration=123.684789486 podStartE2EDuration="2m3.684789486s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:11.683843868 +0000 UTC m=+144.281511670" watchObservedRunningTime="2025-11-24 17:29:11.684789486 +0000 UTC m=+144.282457288" Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.704308 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:11 crc kubenswrapper[4808]: E1124 17:29:11.704769 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:12.204754421 +0000 UTC m=+144.802422223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.712419 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" podStartSLOduration=123.712398798 podStartE2EDuration="2m3.712398798s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:11.711299195 +0000 UTC m=+144.308966997" watchObservedRunningTime="2025-11-24 17:29:11.712398798 +0000 UTC m=+144.310066600" Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.788008 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-d5bl6" podStartSLOduration=123.787989879 podStartE2EDuration="2m3.787989879s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:11.752921205 +0000 UTC m=+144.350589007" watchObservedRunningTime="2025-11-24 17:29:11.787989879 +0000 UTC m=+144.385657681" Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.805552 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:11 crc kubenswrapper[4808]: E1124 17:29:11.806228 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:12.306204231 +0000 UTC m=+144.903872033 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.806439 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" event={"ID":"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c","Type":"ContainerStarted","Data":"377c5d376092e2664f222e363786773234422e1146dd2c0bece4b879c8cbfd72"} Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.807960 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-qlpqd" event={"ID":"f3a1a3f1-05ca-4f24-9836-cc831fcad234","Type":"ContainerStarted","Data":"38d7381622112edcd7cd1e158e894a4428699328f6cea9f887a5313609b17b8a"} Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.813827 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" event={"ID":"d0d60e28-fafc-4490-802e-ee578094c9a3","Type":"ContainerStarted","Data":"712a8df6c81d7842224d96a54f191a746ae99a1106225c41bea4f65c941cff5f"} Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.825854 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" event={"ID":"13575d67-6fe6-4ba5-991b-17ef3b2d3878","Type":"ContainerStarted","Data":"aa9898ded42b05feb101a3bee544692886e8e302f4e9c5f2a616738e333e4a27"} Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.863613 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7sbnn" event={"ID":"ab4ee277-dff6-4411-9fe1-72e56da6244b","Type":"ContainerStarted","Data":"5b4bcf36e03df83848d00f29ddbcf1611bc52531d2cda71732989e2025bac45b"} Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.907051 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:11 crc kubenswrapper[4808]: E1124 17:29:11.907444 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:12.407432226 +0000 UTC m=+145.005100028 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.916088 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ct2bv" podStartSLOduration=123.916060753 podStartE2EDuration="2m3.916060753s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:11.913140826 +0000 UTC m=+144.510808628" watchObservedRunningTime="2025-11-24 17:29:11.916060753 +0000 UTC m=+144.513728545" Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.939722 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f" event={"ID":"df575f42-a7d3-42d8-9d03-f5334bbbc35b","Type":"ContainerStarted","Data":"4272a392cb085f6575dcabee6f8a8276d69657750fc632b13c25f41032f4c32f"} Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.957381 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6"] Nov 24 17:29:11 crc kubenswrapper[4808]: I1124 17:29:11.977267 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" event={"ID":"c908f735-bb61-4a7c-a9be-4f354dfe332b","Type":"ContainerStarted","Data":"7dd899d9f713e72afded2d07af2e71b7ff0517ca0646db936b1ccdc6a3b69770"} Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.001300 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp"] Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.031078 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd" event={"ID":"e9c6772a-c49e-441b-86a3-b76d1ea62823","Type":"ContainerStarted","Data":"182c930cf8146b03212295f02bb2b66f1be46fd83a2d06ecfc06649ca4e0a194"} Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.035832 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:12 crc kubenswrapper[4808]: E1124 17:29:12.036426 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:12.536405197 +0000 UTC m=+145.134072999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.071792 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" podStartSLOduration=125.071761249 podStartE2EDuration="2m5.071761249s" podCreationTimestamp="2025-11-24 17:27:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:12.025402179 +0000 UTC m=+144.623069981" watchObservedRunningTime="2025-11-24 17:29:12.071761249 +0000 UTC m=+144.669429051" Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.078898 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-2d6vb"] Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.137266 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:12 crc kubenswrapper[4808]: E1124 17:29:12.137610 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:12.63759587 +0000 UTC m=+145.235263672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.143516 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5c5r4"] Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.147762 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mmrgn" event={"ID":"3a54581b-46f9-4453-b3c0-9cb55e3410f6","Type":"ContainerStarted","Data":"6ae2f8a4bc5e9a7d0d66db3511cde3586a63c51574c0e91949dc34852ba518b3"} Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.165593 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2dclt" podStartSLOduration=125.165561813 podStartE2EDuration="2m5.165561813s" podCreationTimestamp="2025-11-24 17:27:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:12.165352636 +0000 UTC m=+144.763020468" watchObservedRunningTime="2025-11-24 17:29:12.165561813 +0000 UTC m=+144.763229615" Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.184886 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.202110 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-77zbq" Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.204375 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-ns4rm" podStartSLOduration=125.204358648 podStartE2EDuration="2m5.204358648s" podCreationTimestamp="2025-11-24 17:27:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:12.201683358 +0000 UTC m=+144.799351170" watchObservedRunningTime="2025-11-24 17:29:12.204358648 +0000 UTC m=+144.802026440" Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.208763 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2v2lb"] Nov 24 17:29:12 crc kubenswrapper[4808]: W1124 17:29:12.224462 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod932ea52a_d515_4a68_b53a_84864d62fb37.slice/crio-991ec83d53f10744fcc960cd0419c138c9dac8e00e3e8805501133041cd32cc5 WatchSource:0}: Error finding container 991ec83d53f10744fcc960cd0419c138c9dac8e00e3e8805501133041cd32cc5: Status 404 returned error can't find the container with id 991ec83d53f10744fcc960cd0419c138c9dac8e00e3e8805501133041cd32cc5 Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.240920 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:12 crc kubenswrapper[4808]: E1124 17:29:12.242161 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:12.741977478 +0000 UTC m=+145.339645290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.267065 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc"] Nov 24 17:29:12 crc kubenswrapper[4808]: W1124 17:29:12.331376 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94a29bbc_ed8f_46cf_9917_2ccba227dbcb.slice/crio-911ee731b9bbf3097b739630fdd1d77b3004d27a667b87dd83defa39d86966dc WatchSource:0}: Error finding container 911ee731b9bbf3097b739630fdd1d77b3004d27a667b87dd83defa39d86966dc: Status 404 returned error can't find the container with id 911ee731b9bbf3097b739630fdd1d77b3004d27a667b87dd83defa39d86966dc Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.346656 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:12 crc kubenswrapper[4808]: E1124 17:29:12.347084 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:12.847069718 +0000 UTC m=+145.444737520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.449215 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:12 crc kubenswrapper[4808]: E1124 17:29:12.449660 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:12.949643212 +0000 UTC m=+145.547311014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.476531 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" podStartSLOduration=125.476503322 podStartE2EDuration="2m5.476503322s" podCreationTimestamp="2025-11-24 17:27:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:12.448165798 +0000 UTC m=+145.045833620" watchObservedRunningTime="2025-11-24 17:29:12.476503322 +0000 UTC m=+145.074171124" Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.478102 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-nvgpx"] Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.550775 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:12 crc kubenswrapper[4808]: E1124 17:29:12.551237 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:13.051224356 +0000 UTC m=+145.648892158 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.677228 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:12 crc kubenswrapper[4808]: E1124 17:29:12.677731 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:13.177713513 +0000 UTC m=+145.775381315 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.691145 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tls4q"] Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.691835 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wzvvt" podStartSLOduration=125.691801562 podStartE2EDuration="2m5.691801562s" podCreationTimestamp="2025-11-24 17:27:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:12.671983112 +0000 UTC m=+145.269650914" watchObservedRunningTime="2025-11-24 17:29:12.691801562 +0000 UTC m=+145.289469354" Nov 24 17:29:12 crc kubenswrapper[4808]: W1124 17:29:12.698966 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9e3ed7b_d5c9_4d92_9c13_2ceba9741467.slice/crio-1b799a781b4bdd9e5a8ac6ada440e11ba58b608e262d71fbb0447c5e2046004f WatchSource:0}: Error finding container 1b799a781b4bdd9e5a8ac6ada440e11ba58b608e262d71fbb0447c5e2046004f: Status 404 returned error can't find the container with id 1b799a781b4bdd9e5a8ac6ada440e11ba58b608e262d71fbb0447c5e2046004f Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.722815 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wvj5k"] Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.777855 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp"] Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.778731 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:12 crc kubenswrapper[4808]: E1124 17:29:12.779237 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:13.279221936 +0000 UTC m=+145.876889738 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:12 crc kubenswrapper[4808]: W1124 17:29:12.788783 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24aea72e_3aea_48a5_ac6f_8ab8bb516b18.slice/crio-9146be0f0b2ebf2153d51d5d27ffd880f57c1bc23eb59a7318e9bfe042976807 WatchSource:0}: Error finding container 9146be0f0b2ebf2153d51d5d27ffd880f57c1bc23eb59a7318e9bfe042976807: Status 404 returned error can't find the container with id 9146be0f0b2ebf2153d51d5d27ffd880f57c1bc23eb59a7318e9bfe042976807 Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.879488 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:12 crc kubenswrapper[4808]: E1124 17:29:12.879854 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:13.379838442 +0000 UTC m=+145.977506244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:12 crc kubenswrapper[4808]: I1124 17:29:12.985301 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:12 crc kubenswrapper[4808]: E1124 17:29:12.986715 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:13.486690054 +0000 UTC m=+146.084357856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.082096 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-m57fw"] Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.087083 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:13 crc kubenswrapper[4808]: E1124 17:29:13.087458 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:13.587440624 +0000 UTC m=+146.185108426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.088595 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq"] Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.108499 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ljn6r"] Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.120218 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-6xnwl" podStartSLOduration=125.12020204 podStartE2EDuration="2m5.12020204s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:13.11753134 +0000 UTC m=+145.715199132" watchObservedRunningTime="2025-11-24 17:29:13.12020204 +0000 UTC m=+145.717869832" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.140186 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vkxbx"] Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.164243 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-brt82" podStartSLOduration=125.164222741 podStartE2EDuration="2m5.164222741s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:13.16218374 +0000 UTC m=+145.759851542" watchObservedRunningTime="2025-11-24 17:29:13.164222741 +0000 UTC m=+145.761890543" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.178767 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk"] Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.188967 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:13 crc kubenswrapper[4808]: E1124 17:29:13.189434 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:13.689418911 +0000 UTC m=+146.287086713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:13 crc kubenswrapper[4808]: W1124 17:29:13.196980 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6faa23fa_42a4_499b_99b2_b02585cdf38f.slice/crio-487ad8f70c0bd4d13b8b8486f60ce9ce83c750a450b3b4ac06abc50937190100 WatchSource:0}: Error finding container 487ad8f70c0bd4d13b8b8486f60ce9ce83c750a450b3b4ac06abc50937190100: Status 404 returned error can't find the container with id 487ad8f70c0bd4d13b8b8486f60ce9ce83c750a450b3b4ac06abc50937190100 Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.239058 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr"] Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.247089 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" event={"ID":"13575d67-6fe6-4ba5-991b-17ef3b2d3878","Type":"ContainerStarted","Data":"004e2bdf26f9bfb0952529e3b563122193263bc60851351cfed83a70bf3d90b7"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.248280 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.250569 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" event={"ID":"c908f735-bb61-4a7c-a9be-4f354dfe332b","Type":"ContainerStarted","Data":"82cd25b47f30c56a8fb99b6c60317d6ae719d4fa0f9bbf6c94ad39bcf1f7db31"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.251778 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.253924 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-z2qxd" podStartSLOduration=125.253908431 podStartE2EDuration="2m5.253908431s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:13.238671217 +0000 UTC m=+145.836339019" watchObservedRunningTime="2025-11-24 17:29:13.253908431 +0000 UTC m=+145.851576333" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.272503 4808 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-lpbwg container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.272924 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" podUID="c908f735-bb61-4a7c-a9be-4f354dfe332b" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.272774 4808 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-gx962 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" start-of-body= Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.273039 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" podUID="13575d67-6fe6-4ba5-991b-17ef3b2d3878" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" Nov 24 17:29:13 crc kubenswrapper[4808]: E1124 17:29:13.293241 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:13.793216502 +0000 UTC m=+146.390884294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.293102 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.293937 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:13 crc kubenswrapper[4808]: E1124 17:29:13.301908 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:13.801855689 +0000 UTC m=+146.399523491 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:13 crc kubenswrapper[4808]: W1124 17:29:13.332100 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42e9945a_6923_4ada_b2f3_8aaaa3f853fd.slice/crio-2d61d13613cd680b4d3413f6fb2606737d559b7d2e0ad1a67264584c5ee57a06 WatchSource:0}: Error finding container 2d61d13613cd680b4d3413f6fb2606737d559b7d2e0ad1a67264584c5ee57a06: Status 404 returned error can't find the container with id 2d61d13613cd680b4d3413f6fb2606737d559b7d2e0ad1a67264584c5ee57a06 Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.334508 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd" event={"ID":"e9c6772a-c49e-441b-86a3-b76d1ea62823","Type":"ContainerStarted","Data":"545785f045b7e2cec0d277b9870926632d7f6871014e36c52ffbf4f7bb7c870a"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.390662 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp" event={"ID":"932ea52a-d515-4a68-b53a-84864d62fb37","Type":"ContainerStarted","Data":"46b9b9e199cc4e661d439de33f064b9d7680a7a6f40292fc184fff2663e96594"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.390728 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp" event={"ID":"932ea52a-d515-4a68-b53a-84864d62fb37","Type":"ContainerStarted","Data":"991ec83d53f10744fcc960cd0419c138c9dac8e00e3e8805501133041cd32cc5"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.395065 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:13 crc kubenswrapper[4808]: E1124 17:29:13.396220 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:13.896201838 +0000 UTC m=+146.493869640 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.400592 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.400637 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.403233 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nvgpx" event={"ID":"a9e3ed7b-d5c9-4d92-9c13-2ceba9741467","Type":"ContainerStarted","Data":"1b799a781b4bdd9e5a8ac6ada440e11ba58b608e262d71fbb0447c5e2046004f"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.439771 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" podStartSLOduration=125.439729085 podStartE2EDuration="2m5.439729085s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:13.436144378 +0000 UTC m=+146.033812190" watchObservedRunningTime="2025-11-24 17:29:13.439729085 +0000 UTC m=+146.037396887" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.440600 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg"] Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.440629 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5c5r4" event={"ID":"94a29bbc-ed8f-46cf-9917-2ccba227dbcb","Type":"ContainerStarted","Data":"81438866721a4051a6c93fe64f7f912daef2576ca68fb1e5e72230a7c043025a"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.440647 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5c5r4" event={"ID":"94a29bbc-ed8f-46cf-9917-2ccba227dbcb","Type":"ContainerStarted","Data":"911ee731b9bbf3097b739630fdd1d77b3004d27a667b87dd83defa39d86966dc"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.455394 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tls4q" event={"ID":"24aea72e-3aea-48a5-ac6f-8ab8bb516b18","Type":"ContainerStarted","Data":"9146be0f0b2ebf2153d51d5d27ffd880f57c1bc23eb59a7318e9bfe042976807"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.459052 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6" event={"ID":"d594e8b8-8eb4-4cfa-b978-6fa66ca4c379","Type":"ContainerStarted","Data":"0cf1c7d34662fb53878cbcd1915c71c13f6c20e600b241c11211bce9a4f532ec"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.459124 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6" event={"ID":"d594e8b8-8eb4-4cfa-b978-6fa66ca4c379","Type":"ContainerStarted","Data":"95b38e9b29cc5dacd5d1d8cfbc924d4ca44e3fa76df0ac93f2ab354cf7cc9d0e"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.491815 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc" event={"ID":"29292c32-1a21-415c-83ce-1f526a0264b5","Type":"ContainerStarted","Data":"feceafd45df099d3e64baf8c920fc6852d81ca0475025b5f39dce493f65344ca"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.498646 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:13 crc kubenswrapper[4808]: E1124 17:29:13.500861 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:14.000850765 +0000 UTC m=+146.598518567 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.515882 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chwdd" podStartSLOduration=125.515859232 podStartE2EDuration="2m5.515859232s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:13.471508271 +0000 UTC m=+146.069176093" watchObservedRunningTime="2025-11-24 17:29:13.515859232 +0000 UTC m=+146.113527034" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.516785 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" podStartSLOduration=125.516779819 podStartE2EDuration="2m5.516779819s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:13.514415529 +0000 UTC m=+146.112083321" watchObservedRunningTime="2025-11-24 17:29:13.516779819 +0000 UTC m=+146.114447621" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.534502 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7sbnn" event={"ID":"ab4ee277-dff6-4411-9fe1-72e56da6244b","Type":"ContainerStarted","Data":"9779d921443aeadcc49c482c8fa63dd3050c0ac72175eaf2a6088e3444169c4a"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.542563 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-b4lwp" podStartSLOduration=125.542544686 podStartE2EDuration="2m5.542544686s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:13.541256458 +0000 UTC m=+146.138924260" watchObservedRunningTime="2025-11-24 17:29:13.542544686 +0000 UTC m=+146.140212488" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.555213 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2d6vb" event={"ID":"1b489c90-5e8b-4a49-8505-a25fc32b45c8","Type":"ContainerStarted","Data":"862a6612201d4632530003055fab4e0f6fdd57b5afa2fef2c19a4d1f73e770dd"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.556295 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" event={"ID":"22764299-441e-4ac3-986a-5a4c1a325816","Type":"ContainerStarted","Data":"1a5059cb5e8cc8585c31b2d462b101a3ba42f14574907679f0a916510db3679c"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.562858 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mmrgn" event={"ID":"3a54581b-46f9-4453-b3c0-9cb55e3410f6","Type":"ContainerStarted","Data":"b47a3ff0e1a5ce6ff4f0ad68ae7b115f078e925510bac80737e10cea6d7c1602"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.575934 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-qlpqd" event={"ID":"f3a1a3f1-05ca-4f24-9836-cc831fcad234","Type":"ContainerStarted","Data":"5f20d1dc269ac9d1a12f98715aec66557ee5482ab3623b520f1b4484d01e7518"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.584914 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-5c5r4" podStartSLOduration=125.584900518 podStartE2EDuration="2m5.584900518s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:13.584401373 +0000 UTC m=+146.182069175" watchObservedRunningTime="2025-11-24 17:29:13.584900518 +0000 UTC m=+146.182568320" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.591644 4808 generic.go:334] "Generic (PLEG): container finished" podID="59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c" containerID="d10095ab9dd8f51ea8b983670a9d09deb3966db2c94626c37495ab6e5a76b3fc" exitCode=0 Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.591716 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" event={"ID":"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c","Type":"ContainerDied","Data":"d10095ab9dd8f51ea8b983670a9d09deb3966db2c94626c37495ab6e5a76b3fc"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.603680 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:13 crc kubenswrapper[4808]: E1124 17:29:13.604144 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:14.10412258 +0000 UTC m=+146.701790382 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.610688 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" event={"ID":"664e0a86-ebcc-4119-948c-feefad6325a6","Type":"ContainerStarted","Data":"98d0f0bf5e292dff48a984d9f9b369969de9a14fefade2000ccaadabfba07c71"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.643073 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" event={"ID":"d0d60e28-fafc-4490-802e-ee578094c9a3","Type":"ContainerStarted","Data":"19e5c8ff96219b9fb2f9b99395f0296e89569dc4a1d45586980af9ef69cdc61a"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.648495 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" event={"ID":"94a5dac5-a893-481f-ab4e-bfbd4106b8c6","Type":"ContainerStarted","Data":"3799a0fbac72e744e334d592b8c23c266390e3b8efe6eab592facaf0d60c4dda"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.649276 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.653056 4808 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2v2lb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.653120 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" podUID="94a5dac5-a893-481f-ab4e-bfbd4106b8c6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.655111 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-mmrgn" podStartSLOduration=125.655086498 podStartE2EDuration="2m5.655086498s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:13.630411543 +0000 UTC m=+146.228079355" watchObservedRunningTime="2025-11-24 17:29:13.655086498 +0000 UTC m=+146.252754300" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.657638 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-7sbnn" podStartSLOduration=6.657618373 podStartE2EDuration="6.657618373s" podCreationTimestamp="2025-11-24 17:29:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:13.653730177 +0000 UTC m=+146.251397979" watchObservedRunningTime="2025-11-24 17:29:13.657618373 +0000 UTC m=+146.255286175" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.660435 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f" event={"ID":"df575f42-a7d3-42d8-9d03-f5334bbbc35b","Type":"ContainerStarted","Data":"a7edbe606ce8a5c70f4e602fc53a4f3e1633d5299200042092e771be8fa5eec2"} Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.708181 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:13 crc kubenswrapper[4808]: E1124 17:29:13.708807 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:14.208793847 +0000 UTC m=+146.806461649 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.759675 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" podStartSLOduration=125.759656572 podStartE2EDuration="2m5.759656572s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:13.747441718 +0000 UTC m=+146.345109520" watchObservedRunningTime="2025-11-24 17:29:13.759656572 +0000 UTC m=+146.357324374" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.817693 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:13 crc kubenswrapper[4808]: E1124 17:29:13.818194 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:14.318176014 +0000 UTC m=+146.915843816 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.823247 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfd6f" podStartSLOduration=125.823227895 podStartE2EDuration="2m5.823227895s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:13.821973197 +0000 UTC m=+146.419640989" watchObservedRunningTime="2025-11-24 17:29:13.823227895 +0000 UTC m=+146.420895687" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.824171 4808 patch_prober.go:28] interesting pod/apiserver-76f77b778f-8wq8w container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 24 17:29:13 crc kubenswrapper[4808]: [+]log ok Nov 24 17:29:13 crc kubenswrapper[4808]: [+]etcd ok Nov 24 17:29:13 crc kubenswrapper[4808]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 24 17:29:13 crc kubenswrapper[4808]: [+]poststarthook/generic-apiserver-start-informers ok Nov 24 17:29:13 crc kubenswrapper[4808]: [+]poststarthook/max-in-flight-filter ok Nov 24 17:29:13 crc kubenswrapper[4808]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 24 17:29:13 crc kubenswrapper[4808]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 24 17:29:13 crc kubenswrapper[4808]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 24 17:29:13 crc kubenswrapper[4808]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Nov 24 17:29:13 crc kubenswrapper[4808]: [+]poststarthook/project.openshift.io-projectcache ok Nov 24 17:29:13 crc kubenswrapper[4808]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 24 17:29:13 crc kubenswrapper[4808]: [+]poststarthook/openshift.io-startinformers ok Nov 24 17:29:13 crc kubenswrapper[4808]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 24 17:29:13 crc kubenswrapper[4808]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 24 17:29:13 crc kubenswrapper[4808]: livez check failed Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.824220 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" podUID="8b35999c-bfc9-44bd-bca5-a97e645a7ed9" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.824992 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" podStartSLOduration=125.824985217 podStartE2EDuration="2m5.824985217s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:13.785510862 +0000 UTC m=+146.383178694" watchObservedRunningTime="2025-11-24 17:29:13.824985217 +0000 UTC m=+146.422653019" Nov 24 17:29:13 crc kubenswrapper[4808]: I1124 17:29:13.919157 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:13 crc kubenswrapper[4808]: E1124 17:29:13.919618 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:14.419603165 +0000 UTC m=+147.017270967 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.020237 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:14 crc kubenswrapper[4808]: E1124 17:29:14.020785 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:14.520726246 +0000 UTC m=+147.118394048 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.020965 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:14 crc kubenswrapper[4808]: E1124 17:29:14.021403 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:14.521383946 +0000 UTC m=+147.119051748 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.122175 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:14 crc kubenswrapper[4808]: E1124 17:29:14.122564 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:14.622545938 +0000 UTC m=+147.220213730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.227251 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:14 crc kubenswrapper[4808]: E1124 17:29:14.227691 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:14.727675919 +0000 UTC m=+147.325343721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.328545 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:14 crc kubenswrapper[4808]: E1124 17:29:14.329257 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:14.829221213 +0000 UTC m=+147.426889005 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.433432 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:14 crc kubenswrapper[4808]: E1124 17:29:14.433784 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:14.933770736 +0000 UTC m=+147.531438538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.531126 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.534960 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:14 crc kubenswrapper[4808]: E1124 17:29:14.535426 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:15.035408862 +0000 UTC m=+147.633076664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.548276 4808 patch_prober.go:28] interesting pod/router-default-5444994796-mmrgn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:29:14 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Nov 24 17:29:14 crc kubenswrapper[4808]: [+]process-running ok Nov 24 17:29:14 crc kubenswrapper[4808]: healthz check failed Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.548343 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmrgn" podUID="3a54581b-46f9-4453-b3c0-9cb55e3410f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.637986 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:14 crc kubenswrapper[4808]: E1124 17:29:14.638425 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:15.13840986 +0000 UTC m=+147.736077672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.739765 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:14 crc kubenswrapper[4808]: E1124 17:29:14.740285 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:15.240264353 +0000 UTC m=+147.837932155 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.758296 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc" event={"ID":"29292c32-1a21-415c-83ce-1f526a0264b5","Type":"ContainerStarted","Data":"6afef51d18b0c52e4886b1243eaa615c6e8a4640c1349b5669586e509618a33a"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.771130 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tls4q" event={"ID":"24aea72e-3aea-48a5-ac6f-8ab8bb516b18","Type":"ContainerStarted","Data":"ec412a0d3dfca1863f29ca75de876b7dc64d814c223b90a30918db6b17534451"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.788711 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" event={"ID":"22764299-441e-4ac3-986a-5a4c1a325816","Type":"ContainerStarted","Data":"b5388cb49953f64587bbfca4fb3d8d8c02429379cafdbbce3264610cddef813a"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.788758 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" event={"ID":"22764299-441e-4ac3-986a-5a4c1a325816","Type":"ContainerStarted","Data":"4914c611c51f863acb559f9105adfb688fb41eb5df2898fc8a974585ad329fde"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.790941 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk" event={"ID":"72380442-6dd0-45fd-82ce-07191fb9427e","Type":"ContainerStarted","Data":"dbd1e522cafc7b08f92cea27a34e7b34420b28ea1de58101e91fff4bc3432ec5"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.791005 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk" event={"ID":"72380442-6dd0-45fd-82ce-07191fb9427e","Type":"ContainerStarted","Data":"b3504980f8393613396ce2bc87b97754c108ab4c911b92f6618f70467ee53499"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.810477 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ljn6r" event={"ID":"4319fe26-e7cc-4665-881a-6275268b6330","Type":"ContainerStarted","Data":"203e5cdfdb645fabd027fbd59c1233a1bf60c9471c6773844e11a702838a0378"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.810531 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ljn6r" event={"ID":"4319fe26-e7cc-4665-881a-6275268b6330","Type":"ContainerStarted","Data":"0e0f8dab1b56fccfe7af3036f37bb8f9829757b9a46b619c00d1dea05de69c62"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.813831 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgrtc" podStartSLOduration=126.813809153 podStartE2EDuration="2m6.813809153s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:14.812418681 +0000 UTC m=+147.410086483" watchObservedRunningTime="2025-11-24 17:29:14.813809153 +0000 UTC m=+147.411476955" Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.836203 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" event={"ID":"0706d14f-f31b-4935-a1f5-6e1af9f11b13","Type":"ContainerStarted","Data":"5958e15760ff76097fce945cb6ef57b9237885da9ef1031add677f67fc8c8125"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.841136 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:14 crc kubenswrapper[4808]: E1124 17:29:14.841537 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:15.341519828 +0000 UTC m=+147.939187630 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.846846 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" event={"ID":"42e9945a-6923-4ada-b2f3-8aaaa3f853fd","Type":"ContainerStarted","Data":"e6d49af56897c2ce40c88ddcaf8c98c198d47c85ca1077c3564373164fe54cbc"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.846900 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" event={"ID":"42e9945a-6923-4ada-b2f3-8aaaa3f853fd","Type":"ContainerStarted","Data":"2d61d13613cd680b4d3413f6fb2606737d559b7d2e0ad1a67264584c5ee57a06"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.849139 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6" event={"ID":"d594e8b8-8eb4-4cfa-b978-6fa66ca4c379","Type":"ContainerStarted","Data":"21ab207cfd5e80952754fec1d015985fd5f6055ce00af677db9cd89a104696e7"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.849235 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6" Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.872512 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-qlpqd" event={"ID":"f3a1a3f1-05ca-4f24-9836-cc831fcad234","Type":"ContainerStarted","Data":"79dc4fe48dfb1eb7323dbcba237d5ae5d52279b93eeea80043c1b2d5fd26ce09"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.889624 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-m57fw" event={"ID":"6faa23fa-42a4-499b-99b2-b02585cdf38f","Type":"ContainerStarted","Data":"7fcba3615037ecece729fa96d24ac17359eb9a25faa706b5b9d4c22f55290fd5"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.889677 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-m57fw" event={"ID":"6faa23fa-42a4-499b-99b2-b02585cdf38f","Type":"ContainerStarted","Data":"487ad8f70c0bd4d13b8b8486f60ce9ce83c750a450b3b4ac06abc50937190100"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.892858 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" event={"ID":"4c27c616-6ab4-40fb-857f-cfd4d507547e","Type":"ContainerStarted","Data":"b2fba1965880ebdd2cdeadcc29f46ae69c1c7d0b678b5554500e98f1b1dace5e"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.892905 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" event={"ID":"4c27c616-6ab4-40fb-857f-cfd4d507547e","Type":"ContainerStarted","Data":"97333df0cc117bb91aeef4776c5e977c3295db730e90face98527a9f186c8bb3"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.902450 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.909071 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2d6vb" event={"ID":"1b489c90-5e8b-4a49-8505-a25fc32b45c8","Type":"ContainerStarted","Data":"16d81b38362f55b4f9e28ce9b97e7fbc0b1dada14c2b987a2e1dcb610fc97020"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.911719 4808 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-xlpkq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.911778 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" podUID="4c27c616-6ab4-40fb-857f-cfd4d507547e" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.915914 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr" event={"ID":"81133ff7-9de8-4958-b47c-7513df033b38","Type":"ContainerStarted","Data":"ee795b88d65f98c8b067e427f2e9024f4b15d76c71e3e18772bf01d325671382"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.915975 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr" event={"ID":"81133ff7-9de8-4958-b47c-7513df033b38","Type":"ContainerStarted","Data":"07717cfa8e893f1ac88c0f1946e7b846f580954f3cf2119444d70d8432341a28"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.919109 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nvgpx" event={"ID":"a9e3ed7b-d5c9-4d92-9c13-2ceba9741467","Type":"ContainerStarted","Data":"c49918663b17c547bbc61f0592d7df85c42d382eb73331834f1de03ae9c0cb83"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.919201 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nvgpx" event={"ID":"a9e3ed7b-d5c9-4d92-9c13-2ceba9741467","Type":"ContainerStarted","Data":"5e48634f2a462aabdbb8a2be2c9cb88630dc3d3759477abd3677a309bd47aa2e"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.944286 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-w2bpk" podStartSLOduration=126.944268838 podStartE2EDuration="2m6.944268838s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:14.886415835 +0000 UTC m=+147.484083637" watchObservedRunningTime="2025-11-24 17:29:14.944268838 +0000 UTC m=+147.541936640" Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.944812 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ljn6r" podStartSLOduration=126.944806164 podStartE2EDuration="2m6.944806164s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:14.9359344 +0000 UTC m=+147.533602202" watchObservedRunningTime="2025-11-24 17:29:14.944806164 +0000 UTC m=+147.542473986" Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.946161 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:14 crc kubenswrapper[4808]: E1124 17:29:14.947648 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:15.447628978 +0000 UTC m=+148.045296780 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.970666 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" event={"ID":"94a5dac5-a893-481f-ab4e-bfbd4106b8c6","Type":"ContainerStarted","Data":"8b1591ff5b64fd5ed0c2ae4b6e0e871c54201749042134594259363a5cf31f3e"} Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.976387 4808 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2v2lb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.976460 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" podUID="94a5dac5-a893-481f-ab4e-bfbd4106b8c6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 24 17:29:14 crc kubenswrapper[4808]: I1124 17:29:14.993336 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpbwg" Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.007947 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c4svp" podStartSLOduration=127.007927013 podStartE2EDuration="2m7.007927013s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:15.007658035 +0000 UTC m=+147.605325837" watchObservedRunningTime="2025-11-24 17:29:15.007927013 +0000 UTC m=+147.605594815" Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.047848 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:15 crc kubenswrapper[4808]: E1124 17:29:15.049707 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:15.549695247 +0000 UTC m=+148.147363039 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.094288 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-qlpqd" podStartSLOduration=127.094262594 podStartE2EDuration="2m7.094262594s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:15.090946026 +0000 UTC m=+147.688613828" watchObservedRunningTime="2025-11-24 17:29:15.094262594 +0000 UTC m=+147.691930396" Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.153088 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:15 crc kubenswrapper[4808]: E1124 17:29:15.153379 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:15.653360294 +0000 UTC m=+148.251028096 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.178888 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" podStartSLOduration=127.178867884 podStartE2EDuration="2m7.178867884s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:15.176335168 +0000 UTC m=+147.774002970" watchObservedRunningTime="2025-11-24 17:29:15.178867884 +0000 UTC m=+147.776535686" Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.248347 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6" podStartSLOduration=127.248326702 podStartE2EDuration="2m7.248326702s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:15.24354163 +0000 UTC m=+147.841209432" watchObservedRunningTime="2025-11-24 17:29:15.248326702 +0000 UTC m=+147.845994504" Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.254790 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:15 crc kubenswrapper[4808]: E1124 17:29:15.255123 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:15.755111824 +0000 UTC m=+148.352779626 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.266537 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-vkxbx" podStartSLOduration=127.266514064 podStartE2EDuration="2m7.266514064s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:15.263319199 +0000 UTC m=+147.860987001" watchObservedRunningTime="2025-11-24 17:29:15.266514064 +0000 UTC m=+147.864181866" Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.321168 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nvgpx" podStartSLOduration=127.321151141 podStartE2EDuration="2m7.321151141s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:15.320642826 +0000 UTC m=+147.918310628" watchObservedRunningTime="2025-11-24 17:29:15.321151141 +0000 UTC m=+147.918818943" Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.322253 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-m57fw" podStartSLOduration=8.322248934 podStartE2EDuration="8.322248934s" podCreationTimestamp="2025-11-24 17:29:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:15.292270731 +0000 UTC m=+147.889938533" watchObservedRunningTime="2025-11-24 17:29:15.322248934 +0000 UTC m=+147.919916736" Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.365478 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:15 crc kubenswrapper[4808]: E1124 17:29:15.372931 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:15.872862141 +0000 UTC m=+148.470530023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.469197 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:15 crc kubenswrapper[4808]: E1124 17:29:15.469656 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:15.969635922 +0000 UTC m=+148.567303734 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.524960 4808 patch_prober.go:28] interesting pod/router-default-5444994796-mmrgn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:29:15 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Nov 24 17:29:15 crc kubenswrapper[4808]: [+]process-running ok Nov 24 17:29:15 crc kubenswrapper[4808]: healthz check failed Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.525047 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmrgn" podUID="3a54581b-46f9-4453-b3c0-9cb55e3410f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.571687 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:15 crc kubenswrapper[4808]: E1124 17:29:15.572283 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.072254808 +0000 UTC m=+148.669922610 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.673341 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:15 crc kubenswrapper[4808]: E1124 17:29:15.673840 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.173816223 +0000 UTC m=+148.771484025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.775112 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:15 crc kubenswrapper[4808]: E1124 17:29:15.775562 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.275543512 +0000 UTC m=+148.873211314 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.877177 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:15 crc kubenswrapper[4808]: E1124 17:29:15.877634 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.377609911 +0000 UTC m=+148.975277713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.932611 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fwnlt"] Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.934525 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.939304 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.953313 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fwnlt"] Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.973067 4808 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-gx962 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.973158 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" podUID="13575d67-6fe6-4ba5-991b-17ef3b2d3878" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.977290 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tls4q" event={"ID":"24aea72e-3aea-48a5-ac6f-8ab8bb516b18","Type":"ContainerStarted","Data":"1a5652ecf9cf515567fe55182aa539d4b3132a0a66626b045ff71f19c2559f22"} Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.977934 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:15 crc kubenswrapper[4808]: E1124 17:29:15.978161 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.478132045 +0000 UTC m=+149.075799847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.978454 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:15 crc kubenswrapper[4808]: E1124 17:29:15.978835 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.478822685 +0000 UTC m=+149.076490487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.979915 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2d6vb" event={"ID":"1b489c90-5e8b-4a49-8505-a25fc32b45c8","Type":"ContainerStarted","Data":"adf0627ece118140c56c0b95791a6e0a103ca9503b61e39404a8c71aa2156054"} Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.980062 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-2d6vb" Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.981882 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" event={"ID":"0706d14f-f31b-4935-a1f5-6e1af9f11b13","Type":"ContainerStarted","Data":"610890b8397526f6c44d6d7753545a1334d282cd7ec8dd9648fd1b1256b1d02d"} Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.981911 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" event={"ID":"0706d14f-f31b-4935-a1f5-6e1af9f11b13","Type":"ContainerStarted","Data":"208cd192734657ee5b812aeb0de69bcf0af0e9c751f83d9d6f6ef13e7ebfa885"} Nov 24 17:29:15 crc kubenswrapper[4808]: I1124 17:29:15.987609 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr" event={"ID":"81133ff7-9de8-4958-b47c-7513df033b38","Type":"ContainerStarted","Data":"f1ce67f032249278383ee624a5505a8376818c1faeae2ce75690c7efe1d9da20"} Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.003458 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" event={"ID":"59718ce4-1a5a-41e5-b6dd-bd318cf5cb3c","Type":"ContainerStarted","Data":"4ef25ccda0ed3d32315a55bc9b4fcd5b7daac98c80f9ab9d310e9f8df7af00d1"} Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.008479 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" event={"ID":"664e0a86-ebcc-4119-948c-feefad6325a6","Type":"ContainerStarted","Data":"462ca259dda1e3b4867cd4b3f378f46f55bbd89f769e5170ba0226f2f93b5854"} Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.010408 4808 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2v2lb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.010468 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" podUID="94a5dac5-a893-481f-ab4e-bfbd4106b8c6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.011059 4808 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-xlpkq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.011089 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" podUID="4c27c616-6ab4-40fb-857f-cfd4d507547e" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.016173 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gx962" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.019624 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-2d6vb" podStartSLOduration=9.01960277 podStartE2EDuration="9.01960277s" podCreationTimestamp="2025-11-24 17:29:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:16.018123226 +0000 UTC m=+148.615791028" watchObservedRunningTime="2025-11-24 17:29:16.01960277 +0000 UTC m=+148.617270572" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.047195 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-tls4q" podStartSLOduration=128.04716444 podStartE2EDuration="2m8.04716444s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:16.04713451 +0000 UTC m=+148.644802312" watchObservedRunningTime="2025-11-24 17:29:16.04716444 +0000 UTC m=+148.644832242" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.074345 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" podStartSLOduration=128.074317948 podStartE2EDuration="2m8.074317948s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:16.071548426 +0000 UTC m=+148.669216248" watchObservedRunningTime="2025-11-24 17:29:16.074317948 +0000 UTC m=+148.671985750" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.079104 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.079485 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f11d89a2-be43-47fd-ae09-57fcd7df9820-utilities\") pod \"certified-operators-fwnlt\" (UID: \"f11d89a2-be43-47fd-ae09-57fcd7df9820\") " pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.079590 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f11d89a2-be43-47fd-ae09-57fcd7df9820-catalog-content\") pod \"certified-operators-fwnlt\" (UID: \"f11d89a2-be43-47fd-ae09-57fcd7df9820\") " pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.079722 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4g78\" (UniqueName: \"kubernetes.io/projected/f11d89a2-be43-47fd-ae09-57fcd7df9820-kube-api-access-w4g78\") pod \"certified-operators-fwnlt\" (UID: \"f11d89a2-be43-47fd-ae09-57fcd7df9820\") " pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:29:16 crc kubenswrapper[4808]: E1124 17:29:16.084389 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.584355637 +0000 UTC m=+149.182023439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.118072 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mlmvz"] Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.119176 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.121125 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.126647 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zcmdg" podStartSLOduration=128.126623746 podStartE2EDuration="2m8.126623746s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:16.126227184 +0000 UTC m=+148.723894986" watchObservedRunningTime="2025-11-24 17:29:16.126623746 +0000 UTC m=+148.724291548" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.149620 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mlmvz"] Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.170796 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9dqfr" podStartSLOduration=128.17077284 podStartE2EDuration="2m8.17077284s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:16.169194303 +0000 UTC m=+148.766862125" watchObservedRunningTime="2025-11-24 17:29:16.17077284 +0000 UTC m=+148.768440642" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.183035 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4g78\" (UniqueName: \"kubernetes.io/projected/f11d89a2-be43-47fd-ae09-57fcd7df9820-kube-api-access-w4g78\") pod \"certified-operators-fwnlt\" (UID: \"f11d89a2-be43-47fd-ae09-57fcd7df9820\") " pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.183147 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.183420 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f11d89a2-be43-47fd-ae09-57fcd7df9820-utilities\") pod \"certified-operators-fwnlt\" (UID: \"f11d89a2-be43-47fd-ae09-57fcd7df9820\") " pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.183552 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f11d89a2-be43-47fd-ae09-57fcd7df9820-catalog-content\") pod \"certified-operators-fwnlt\" (UID: \"f11d89a2-be43-47fd-ae09-57fcd7df9820\") " pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.184134 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f11d89a2-be43-47fd-ae09-57fcd7df9820-catalog-content\") pod \"certified-operators-fwnlt\" (UID: \"f11d89a2-be43-47fd-ae09-57fcd7df9820\") " pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.184403 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f11d89a2-be43-47fd-ae09-57fcd7df9820-utilities\") pod \"certified-operators-fwnlt\" (UID: \"f11d89a2-be43-47fd-ae09-57fcd7df9820\") " pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:29:16 crc kubenswrapper[4808]: E1124 17:29:16.186993 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.686973803 +0000 UTC m=+149.284641605 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.228180 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4g78\" (UniqueName: \"kubernetes.io/projected/f11d89a2-be43-47fd-ae09-57fcd7df9820-kube-api-access-w4g78\") pod \"certified-operators-fwnlt\" (UID: \"f11d89a2-be43-47fd-ae09-57fcd7df9820\") " pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.252504 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.284427 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.284794 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.284825 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.284857 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.284887 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f2246d-a6bb-463f-9a38-2d7144079eb6-utilities\") pod \"community-operators-mlmvz\" (UID: \"a3f2246d-a6bb-463f-9a38-2d7144079eb6\") " pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.284946 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nblks\" (UniqueName: \"kubernetes.io/projected/a3f2246d-a6bb-463f-9a38-2d7144079eb6-kube-api-access-nblks\") pod \"community-operators-mlmvz\" (UID: \"a3f2246d-a6bb-463f-9a38-2d7144079eb6\") " pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.284968 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.284996 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f2246d-a6bb-463f-9a38-2d7144079eb6-catalog-content\") pod \"community-operators-mlmvz\" (UID: \"a3f2246d-a6bb-463f-9a38-2d7144079eb6\") " pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:29:16 crc kubenswrapper[4808]: E1124 17:29:16.285128 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.785109265 +0000 UTC m=+149.382777067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.285982 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.291370 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.297662 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.297981 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.321922 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j9jpf"] Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.323094 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.390768 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f2246d-a6bb-463f-9a38-2d7144079eb6-catalog-content\") pod \"community-operators-mlmvz\" (UID: \"a3f2246d-a6bb-463f-9a38-2d7144079eb6\") " pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.390887 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f2246d-a6bb-463f-9a38-2d7144079eb6-utilities\") pod \"community-operators-mlmvz\" (UID: \"a3f2246d-a6bb-463f-9a38-2d7144079eb6\") " pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.390926 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.390980 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nblks\" (UniqueName: \"kubernetes.io/projected/a3f2246d-a6bb-463f-9a38-2d7144079eb6-kube-api-access-nblks\") pod \"community-operators-mlmvz\" (UID: \"a3f2246d-a6bb-463f-9a38-2d7144079eb6\") " pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.391896 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f2246d-a6bb-463f-9a38-2d7144079eb6-catalog-content\") pod \"community-operators-mlmvz\" (UID: \"a3f2246d-a6bb-463f-9a38-2d7144079eb6\") " pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.392194 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f2246d-a6bb-463f-9a38-2d7144079eb6-utilities\") pod \"community-operators-mlmvz\" (UID: \"a3f2246d-a6bb-463f-9a38-2d7144079eb6\") " pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:29:16 crc kubenswrapper[4808]: E1124 17:29:16.392587 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.892571745 +0000 UTC m=+149.490239547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.413341 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j9jpf"] Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.431573 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nblks\" (UniqueName: \"kubernetes.io/projected/a3f2246d-a6bb-463f-9a38-2d7144079eb6-kube-api-access-nblks\") pod \"community-operators-mlmvz\" (UID: \"a3f2246d-a6bb-463f-9a38-2d7144079eb6\") " pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.435398 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.472456 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.496845 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.497274 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp88c\" (UniqueName: \"kubernetes.io/projected/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-kube-api-access-wp88c\") pod \"certified-operators-j9jpf\" (UID: \"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3\") " pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.497323 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-utilities\") pod \"certified-operators-j9jpf\" (UID: \"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3\") " pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.497388 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-catalog-content\") pod \"certified-operators-j9jpf\" (UID: \"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3\") " pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:29:16 crc kubenswrapper[4808]: E1124 17:29:16.497535 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:16.99751325 +0000 UTC m=+149.595181052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.512374 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4pwnf"] Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.513473 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.526266 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4pwnf"] Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.537332 4808 patch_prober.go:28] interesting pod/router-default-5444994796-mmrgn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:29:16 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Nov 24 17:29:16 crc kubenswrapper[4808]: [+]process-running ok Nov 24 17:29:16 crc kubenswrapper[4808]: healthz check failed Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.537405 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmrgn" podUID="3a54581b-46f9-4453-b3c0-9cb55e3410f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.570345 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.580353 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.602873 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-catalog-content\") pod \"certified-operators-j9jpf\" (UID: \"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3\") " pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.602935 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.602966 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp88c\" (UniqueName: \"kubernetes.io/projected/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-kube-api-access-wp88c\") pod \"certified-operators-j9jpf\" (UID: \"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3\") " pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.602989 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-utilities\") pod \"certified-operators-j9jpf\" (UID: \"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3\") " pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.603907 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-utilities\") pod \"certified-operators-j9jpf\" (UID: \"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3\") " pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.604165 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-catalog-content\") pod \"certified-operators-j9jpf\" (UID: \"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3\") " pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:29:16 crc kubenswrapper[4808]: E1124 17:29:16.604483 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:17.104467765 +0000 UTC m=+149.702135567 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.670763 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp88c\" (UniqueName: \"kubernetes.io/projected/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-kube-api-access-wp88c\") pod \"certified-operators-j9jpf\" (UID: \"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3\") " pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.704669 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.705057 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a0cb760-04cf-4658-b4ff-1da560789800-catalog-content\") pod \"community-operators-4pwnf\" (UID: \"7a0cb760-04cf-4658-b4ff-1da560789800\") " pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.705144 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a0cb760-04cf-4658-b4ff-1da560789800-utilities\") pod \"community-operators-4pwnf\" (UID: \"7a0cb760-04cf-4658-b4ff-1da560789800\") " pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:29:16 crc kubenswrapper[4808]: E1124 17:29:16.705260 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:17.205222816 +0000 UTC m=+149.802890618 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.705318 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thjx9\" (UniqueName: \"kubernetes.io/projected/7a0cb760-04cf-4658-b4ff-1da560789800-kube-api-access-thjx9\") pod \"community-operators-4pwnf\" (UID: \"7a0cb760-04cf-4658-b4ff-1da560789800\") " pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.809589 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thjx9\" (UniqueName: \"kubernetes.io/projected/7a0cb760-04cf-4658-b4ff-1da560789800-kube-api-access-thjx9\") pod \"community-operators-4pwnf\" (UID: \"7a0cb760-04cf-4658-b4ff-1da560789800\") " pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.809671 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a0cb760-04cf-4658-b4ff-1da560789800-catalog-content\") pod \"community-operators-4pwnf\" (UID: \"7a0cb760-04cf-4658-b4ff-1da560789800\") " pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.809714 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.809739 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a0cb760-04cf-4658-b4ff-1da560789800-utilities\") pod \"community-operators-4pwnf\" (UID: \"7a0cb760-04cf-4658-b4ff-1da560789800\") " pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.810251 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a0cb760-04cf-4658-b4ff-1da560789800-utilities\") pod \"community-operators-4pwnf\" (UID: \"7a0cb760-04cf-4658-b4ff-1da560789800\") " pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.810815 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a0cb760-04cf-4658-b4ff-1da560789800-catalog-content\") pod \"community-operators-4pwnf\" (UID: \"7a0cb760-04cf-4658-b4ff-1da560789800\") " pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:29:16 crc kubenswrapper[4808]: E1124 17:29:16.811323 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:17.311299614 +0000 UTC m=+149.908967416 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.863870 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thjx9\" (UniqueName: \"kubernetes.io/projected/7a0cb760-04cf-4658-b4ff-1da560789800-kube-api-access-thjx9\") pod \"community-operators-4pwnf\" (UID: \"7a0cb760-04cf-4658-b4ff-1da560789800\") " pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.910636 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:16 crc kubenswrapper[4808]: E1124 17:29:16.911189 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:17.411171348 +0000 UTC m=+150.008839150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:16 crc kubenswrapper[4808]: I1124 17:29:16.937707 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.023139 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:17 crc kubenswrapper[4808]: E1124 17:29:17.023594 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:17.523576146 +0000 UTC m=+150.121243948 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.072699 4808 generic.go:334] "Generic (PLEG): container finished" podID="d0d60e28-fafc-4490-802e-ee578094c9a3" containerID="19e5c8ff96219b9fb2f9b99395f0296e89569dc4a1d45586980af9ef69cdc61a" exitCode=0 Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.073896 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" event={"ID":"d0d60e28-fafc-4490-802e-ee578094c9a3","Type":"ContainerDied","Data":"19e5c8ff96219b9fb2f9b99395f0296e89569dc4a1d45586980af9ef69cdc61a"} Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.078554 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" Nov 24 17:29:17 crc kubenswrapper[4808]: W1124 17:29:17.092281 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-ea2d7327d5b0c98713250447e60f04b07e51151debb6dec0d4568d785f3d71b3 WatchSource:0}: Error finding container ea2d7327d5b0c98713250447e60f04b07e51151debb6dec0d4568d785f3d71b3: Status 404 returned error can't find the container with id ea2d7327d5b0c98713250447e60f04b07e51151debb6dec0d4568d785f3d71b3 Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.125175 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:17 crc kubenswrapper[4808]: E1124 17:29:17.125530 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:17.625511571 +0000 UTC m=+150.223179373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.150672 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fwnlt"] Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.154440 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.208870 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mlmvz"] Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.230428 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:17 crc kubenswrapper[4808]: E1124 17:29:17.230797 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:17.730781996 +0000 UTC m=+150.328449798 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.332914 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:17 crc kubenswrapper[4808]: E1124 17:29:17.333368 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:17.83333397 +0000 UTC m=+150.431001772 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.435804 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:17 crc kubenswrapper[4808]: E1124 17:29:17.436732 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:17.936716048 +0000 UTC m=+150.534383850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.536793 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:17 crc kubenswrapper[4808]: E1124 17:29:17.537219 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:18.037196141 +0000 UTC m=+150.634863943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.545694 4808 patch_prober.go:28] interesting pod/router-default-5444994796-mmrgn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:29:17 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Nov 24 17:29:17 crc kubenswrapper[4808]: [+]process-running ok Nov 24 17:29:17 crc kubenswrapper[4808]: healthz check failed Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.545811 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmrgn" podUID="3a54581b-46f9-4453-b3c0-9cb55e3410f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.648121 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:17 crc kubenswrapper[4808]: E1124 17:29:17.648689 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:18.14865291 +0000 UTC m=+150.746320712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.749430 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:17 crc kubenswrapper[4808]: E1124 17:29:17.750395 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:18.250375739 +0000 UTC m=+150.848043531 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.852246 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:17 crc kubenswrapper[4808]: E1124 17:29:17.852732 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:18.352714716 +0000 UTC m=+150.950382518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.911599 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-96cn9"] Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.912814 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.915887 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.945115 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4pwnf"] Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.955881 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:17 crc kubenswrapper[4808]: E1124 17:29:17.956864 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:18.456573309 +0000 UTC m=+151.054241111 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.994849 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-96cn9"] Nov 24 17:29:17 crc kubenswrapper[4808]: I1124 17:29:17.997510 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j9jpf"] Nov 24 17:29:18 crc kubenswrapper[4808]: W1124 17:29:18.013196 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1ef2dab_8530_43cf_a72e_22cb8ac8e1a3.slice/crio-e6780a3abcfb50bb0392c0c4786e833d64d6a91750ec1f9be87573b4b1853014 WatchSource:0}: Error finding container e6780a3abcfb50bb0392c0c4786e833d64d6a91750ec1f9be87573b4b1853014: Status 404 returned error can't find the container with id e6780a3abcfb50bb0392c0c4786e833d64d6a91750ec1f9be87573b4b1853014 Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.060387 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.060475 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2r6w\" (UniqueName: \"kubernetes.io/projected/cde9313c-9316-40dc-8427-a8683ce67b5b-kube-api-access-v2r6w\") pod \"redhat-marketplace-96cn9\" (UID: \"cde9313c-9316-40dc-8427-a8683ce67b5b\") " pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.060526 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cde9313c-9316-40dc-8427-a8683ce67b5b-utilities\") pod \"redhat-marketplace-96cn9\" (UID: \"cde9313c-9316-40dc-8427-a8683ce67b5b\") " pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.060563 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cde9313c-9316-40dc-8427-a8683ce67b5b-catalog-content\") pod \"redhat-marketplace-96cn9\" (UID: \"cde9313c-9316-40dc-8427-a8683ce67b5b\") " pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:29:18 crc kubenswrapper[4808]: E1124 17:29:18.061159 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:18.561125483 +0000 UTC m=+151.158793285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.084401 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" event={"ID":"664e0a86-ebcc-4119-948c-feefad6325a6","Type":"ContainerStarted","Data":"c74d2295f8d4c069fa5f74d3e652f8463a3e921f3efd7e0589c2658ce726ab34"} Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.086860 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"20899522df431fe2decc712051d88060346c448547eea12eb58bca8fb291d381"} Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.086900 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1b5929ac634a9ea93fa52e2467009bac8b8077100a1ba4e5403eb55b05327fde"} Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.099665 4808 generic.go:334] "Generic (PLEG): container finished" podID="f11d89a2-be43-47fd-ae09-57fcd7df9820" containerID="898c3b72ff56e4f287688b51154b00b934a3e1b3f59e0ac1997e09e3160f18e8" exitCode=0 Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.100383 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fwnlt" event={"ID":"f11d89a2-be43-47fd-ae09-57fcd7df9820","Type":"ContainerDied","Data":"898c3b72ff56e4f287688b51154b00b934a3e1b3f59e0ac1997e09e3160f18e8"} Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.100486 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fwnlt" event={"ID":"f11d89a2-be43-47fd-ae09-57fcd7df9820","Type":"ContainerStarted","Data":"58f125d1375deb8d27747d24dbe4796c293d2634ffffc55921e3d616f23e9877"} Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.103902 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.110067 4808 generic.go:334] "Generic (PLEG): container finished" podID="a3f2246d-a6bb-463f-9a38-2d7144079eb6" containerID="64744792fad2fd91ce4c09fe70f2e2d13d7d9c14962852daddadaf8d1d791551" exitCode=0 Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.110165 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlmvz" event={"ID":"a3f2246d-a6bb-463f-9a38-2d7144079eb6","Type":"ContainerDied","Data":"64744792fad2fd91ce4c09fe70f2e2d13d7d9c14962852daddadaf8d1d791551"} Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.110221 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlmvz" event={"ID":"a3f2246d-a6bb-463f-9a38-2d7144079eb6","Type":"ContainerStarted","Data":"3f093bdb719ed3e403f63954a18527a1ceffb807e19e09157dce1bea08862407"} Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.116774 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ca9727636305f8c3f061a4ee96610cbcc83c872ff18359ed1c74f3bcb46dbcf7"} Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.117177 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"149424f2ab80318bb2de498f94893858f16aeaaaa50b13938ea49474387948c5"} Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.117662 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.122028 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j9jpf" event={"ID":"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3","Type":"ContainerStarted","Data":"e6780a3abcfb50bb0392c0c4786e833d64d6a91750ec1f9be87573b4b1853014"} Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.136426 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"23e01e7349006149de9caa8c4be435dad52e0051ef1c3f2c2a167af22f0973fb"} Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.136496 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ea2d7327d5b0c98713250447e60f04b07e51151debb6dec0d4568d785f3d71b3"} Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.156542 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4pwnf" event={"ID":"7a0cb760-04cf-4658-b4ff-1da560789800","Type":"ContainerStarted","Data":"7d6d77f6a8ef7dd3c30c52674215fe24c60d8e5c42bd32e8a678659b2bcdc6ea"} Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.163191 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.163397 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2r6w\" (UniqueName: \"kubernetes.io/projected/cde9313c-9316-40dc-8427-a8683ce67b5b-kube-api-access-v2r6w\") pod \"redhat-marketplace-96cn9\" (UID: \"cde9313c-9316-40dc-8427-a8683ce67b5b\") " pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.163476 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cde9313c-9316-40dc-8427-a8683ce67b5b-utilities\") pod \"redhat-marketplace-96cn9\" (UID: \"cde9313c-9316-40dc-8427-a8683ce67b5b\") " pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.163527 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cde9313c-9316-40dc-8427-a8683ce67b5b-catalog-content\") pod \"redhat-marketplace-96cn9\" (UID: \"cde9313c-9316-40dc-8427-a8683ce67b5b\") " pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:29:18 crc kubenswrapper[4808]: E1124 17:29:18.164295 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:18.664268284 +0000 UTC m=+151.261936086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.166411 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cde9313c-9316-40dc-8427-a8683ce67b5b-utilities\") pod \"redhat-marketplace-96cn9\" (UID: \"cde9313c-9316-40dc-8427-a8683ce67b5b\") " pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.166618 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cde9313c-9316-40dc-8427-a8683ce67b5b-catalog-content\") pod \"redhat-marketplace-96cn9\" (UID: \"cde9313c-9316-40dc-8427-a8683ce67b5b\") " pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.201254 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2r6w\" (UniqueName: \"kubernetes.io/projected/cde9313c-9316-40dc-8427-a8683ce67b5b-kube-api-access-v2r6w\") pod \"redhat-marketplace-96cn9\" (UID: \"cde9313c-9316-40dc-8427-a8683ce67b5b\") " pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.237841 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.264573 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:18 crc kubenswrapper[4808]: E1124 17:29:18.265446 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:18.765411536 +0000 UTC m=+151.363079418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.310709 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v5lxz"] Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.311980 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.328377 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5lxz"] Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.355089 4808 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.368717 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.369148 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04baaaad-0212-49d6-86af-d498e40e8600-catalog-content\") pod \"redhat-marketplace-v5lxz\" (UID: \"04baaaad-0212-49d6-86af-d498e40e8600\") " pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.369211 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04baaaad-0212-49d6-86af-d498e40e8600-utilities\") pod \"redhat-marketplace-v5lxz\" (UID: \"04baaaad-0212-49d6-86af-d498e40e8600\") " pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.369266 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m966\" (UniqueName: \"kubernetes.io/projected/04baaaad-0212-49d6-86af-d498e40e8600-kube-api-access-2m966\") pod \"redhat-marketplace-v5lxz\" (UID: \"04baaaad-0212-49d6-86af-d498e40e8600\") " pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:29:18 crc kubenswrapper[4808]: E1124 17:29:18.369489 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:29:18.869474095 +0000 UTC m=+151.467141897 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.411430 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.418735 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-8wq8w" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.446477 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-z2qxd" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.461208 4808 patch_prober.go:28] interesting pod/downloads-7954f5f757-6xnwl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.461284 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6xnwl" podUID="47734a14-98f7-4be3-89a3-74f5a13ab3e3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.467928 4808 patch_prober.go:28] interesting pod/downloads-7954f5f757-6xnwl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.467991 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6xnwl" podUID="47734a14-98f7-4be3-89a3-74f5a13ab3e3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.486550 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04baaaad-0212-49d6-86af-d498e40e8600-catalog-content\") pod \"redhat-marketplace-v5lxz\" (UID: \"04baaaad-0212-49d6-86af-d498e40e8600\") " pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.486755 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.486800 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04baaaad-0212-49d6-86af-d498e40e8600-utilities\") pod \"redhat-marketplace-v5lxz\" (UID: \"04baaaad-0212-49d6-86af-d498e40e8600\") " pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.487045 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m966\" (UniqueName: \"kubernetes.io/projected/04baaaad-0212-49d6-86af-d498e40e8600-kube-api-access-2m966\") pod \"redhat-marketplace-v5lxz\" (UID: \"04baaaad-0212-49d6-86af-d498e40e8600\") " pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:29:18 crc kubenswrapper[4808]: E1124 17:29:18.487265 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:29:18.987246682 +0000 UTC m=+151.584914484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gjnx8" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.489822 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04baaaad-0212-49d6-86af-d498e40e8600-utilities\") pod \"redhat-marketplace-v5lxz\" (UID: \"04baaaad-0212-49d6-86af-d498e40e8600\") " pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.495847 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04baaaad-0212-49d6-86af-d498e40e8600-catalog-content\") pod \"redhat-marketplace-v5lxz\" (UID: \"04baaaad-0212-49d6-86af-d498e40e8600\") " pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.526538 4808 patch_prober.go:28] interesting pod/router-default-5444994796-mmrgn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:29:18 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Nov 24 17:29:18 crc kubenswrapper[4808]: [+]process-running ok Nov 24 17:29:18 crc kubenswrapper[4808]: healthz check failed Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.526597 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmrgn" podUID="3a54581b-46f9-4453-b3c0-9cb55e3410f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.531373 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.533493 4808 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-24T17:29:18.355120207Z","Handler":null,"Name":""} Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.554087 4808 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.554138 4808 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.554693 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m966\" (UniqueName: \"kubernetes.io/projected/04baaaad-0212-49d6-86af-d498e40e8600-kube-api-access-2m966\") pod \"redhat-marketplace-v5lxz\" (UID: \"04baaaad-0212-49d6-86af-d498e40e8600\") " pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.590148 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0d60e28-fafc-4490-802e-ee578094c9a3-config-volume\") pod \"d0d60e28-fafc-4490-802e-ee578094c9a3\" (UID: \"d0d60e28-fafc-4490-802e-ee578094c9a3\") " Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.590210 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc57x\" (UniqueName: \"kubernetes.io/projected/d0d60e28-fafc-4490-802e-ee578094c9a3-kube-api-access-sc57x\") pod \"d0d60e28-fafc-4490-802e-ee578094c9a3\" (UID: \"d0d60e28-fafc-4490-802e-ee578094c9a3\") " Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.590240 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0d60e28-fafc-4490-802e-ee578094c9a3-secret-volume\") pod \"d0d60e28-fafc-4490-802e-ee578094c9a3\" (UID: \"d0d60e28-fafc-4490-802e-ee578094c9a3\") " Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.590463 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.593374 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0d60e28-fafc-4490-802e-ee578094c9a3-config-volume" (OuterVolumeSpecName: "config-volume") pod "d0d60e28-fafc-4490-802e-ee578094c9a3" (UID: "d0d60e28-fafc-4490-802e-ee578094c9a3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.601028 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0d60e28-fafc-4490-802e-ee578094c9a3-kube-api-access-sc57x" (OuterVolumeSpecName: "kube-api-access-sc57x") pod "d0d60e28-fafc-4490-802e-ee578094c9a3" (UID: "d0d60e28-fafc-4490-802e-ee578094c9a3"). InnerVolumeSpecName "kube-api-access-sc57x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.605313 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d60e28-fafc-4490-802e-ee578094c9a3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d0d60e28-fafc-4490-802e-ee578094c9a3" (UID: "d0d60e28-fafc-4490-802e-ee578094c9a3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.635670 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-96cn9"] Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.635855 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.684444 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.691230 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.691412 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0d60e28-fafc-4490-802e-ee578094c9a3-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.693402 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc57x\" (UniqueName: \"kubernetes.io/projected/d0d60e28-fafc-4490-802e-ee578094c9a3-kube-api-access-sc57x\") on node \"crc\" DevicePath \"\"" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.693444 4808 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0d60e28-fafc-4490-802e-ee578094c9a3-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.696230 4808 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.696282 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.774993 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gjnx8\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.817466 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.818167 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.821496 4808 patch_prober.go:28] interesting pod/console-f9d7485db-brt82 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.821587 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-brt82" podUID="e87a6b3d-2bbf-49a1-9209-a1327c4bd30d" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.877536 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:18 crc kubenswrapper[4808]: I1124 17:29:18.960995 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5lxz"] Nov 24 17:29:18 crc kubenswrapper[4808]: W1124 17:29:18.972589 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04baaaad_0212_49d6_86af_d498e40e8600.slice/crio-03861f27f5b5679fb91dee5069bde6496967ae4ea12c0c1c3b728e50aec95cd7 WatchSource:0}: Error finding container 03861f27f5b5679fb91dee5069bde6496967ae4ea12c0c1c3b728e50aec95cd7: Status 404 returned error can't find the container with id 03861f27f5b5679fb91dee5069bde6496967ae4ea12c0c1c3b728e50aec95cd7 Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.121357 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5wdrd"] Nov 24 17:29:19 crc kubenswrapper[4808]: E1124 17:29:19.121602 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d60e28-fafc-4490-802e-ee578094c9a3" containerName="collect-profiles" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.121618 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d60e28-fafc-4490-802e-ee578094c9a3" containerName="collect-profiles" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.121756 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0d60e28-fafc-4490-802e-ee578094c9a3" containerName="collect-profiles" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.122569 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.128594 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.141724 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5wdrd"] Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.179134 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.179949 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.182837 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.182837 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.188625 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.195486 4808 generic.go:334] "Generic (PLEG): container finished" podID="7a0cb760-04cf-4658-b4ff-1da560789800" containerID="2094c8fbdfa10f3e12bba5a8aaec3e8bee1fd061bdfc4ab9d85dd7f49f972743" exitCode=0 Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.195616 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4pwnf" event={"ID":"7a0cb760-04cf-4658-b4ff-1da560789800","Type":"ContainerDied","Data":"2094c8fbdfa10f3e12bba5a8aaec3e8bee1fd061bdfc4ab9d85dd7f49f972743"} Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.208198 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b005cb77-3bba-44a5-8446-f6fcc5053336-utilities\") pod \"redhat-operators-5wdrd\" (UID: \"b005cb77-3bba-44a5-8446-f6fcc5053336\") " pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.208236 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ea04db8-7c03-4df7-bbd5-914915bdeea0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7ea04db8-7c03-4df7-bbd5-914915bdeea0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.208286 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqn4d\" (UniqueName: \"kubernetes.io/projected/b005cb77-3bba-44a5-8446-f6fcc5053336-kube-api-access-qqn4d\") pod \"redhat-operators-5wdrd\" (UID: \"b005cb77-3bba-44a5-8446-f6fcc5053336\") " pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.208310 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b005cb77-3bba-44a5-8446-f6fcc5053336-catalog-content\") pod \"redhat-operators-5wdrd\" (UID: \"b005cb77-3bba-44a5-8446-f6fcc5053336\") " pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.208330 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ea04db8-7c03-4df7-bbd5-914915bdeea0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7ea04db8-7c03-4df7-bbd5-914915bdeea0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.213840 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" event={"ID":"664e0a86-ebcc-4119-948c-feefad6325a6","Type":"ContainerStarted","Data":"b3f0c5a38c751276f132ea85aa97c16944986d456e5ae0da3d53fb9041571219"} Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.213894 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" event={"ID":"664e0a86-ebcc-4119-948c-feefad6325a6","Type":"ContainerStarted","Data":"99676d2bdfbde517184946a9c76bbcdd08cc6ddf3b5b78d62782dffedabc591d"} Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.244237 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" event={"ID":"d0d60e28-fafc-4490-802e-ee578094c9a3","Type":"ContainerDied","Data":"712a8df6c81d7842224d96a54f191a746ae99a1106225c41bea4f65c941cff5f"} Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.244288 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="712a8df6c81d7842224d96a54f191a746ae99a1106225c41bea4f65c941cff5f" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.244339 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.250034 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gjnx8"] Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.277670 4808 generic.go:334] "Generic (PLEG): container finished" podID="b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" containerID="af9fedb99208e3b636087707a803b50e5a7e32756663b13291a7cc892986231d" exitCode=0 Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.278118 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j9jpf" event={"ID":"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3","Type":"ContainerDied","Data":"af9fedb99208e3b636087707a803b50e5a7e32756663b13291a7cc892986231d"} Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.283851 4808 generic.go:334] "Generic (PLEG): container finished" podID="cde9313c-9316-40dc-8427-a8683ce67b5b" containerID="01542eaca128f2d4d04cb61fdc5f02e4cf7503fca6f0f0de6b39740c2863931b" exitCode=0 Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.283906 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96cn9" event={"ID":"cde9313c-9316-40dc-8427-a8683ce67b5b","Type":"ContainerDied","Data":"01542eaca128f2d4d04cb61fdc5f02e4cf7503fca6f0f0de6b39740c2863931b"} Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.283956 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96cn9" event={"ID":"cde9313c-9316-40dc-8427-a8683ce67b5b","Type":"ContainerStarted","Data":"76a47db264d65b885bec218b6f08f2c697e7ce16475c28980a69fb733df1f546"} Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.287059 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5lxz" event={"ID":"04baaaad-0212-49d6-86af-d498e40e8600","Type":"ContainerStarted","Data":"03861f27f5b5679fb91dee5069bde6496967ae4ea12c0c1c3b728e50aec95cd7"} Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.302799 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-wvj5k" podStartSLOduration=12.302770047 podStartE2EDuration="12.302770047s" podCreationTimestamp="2025-11-24 17:29:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:19.261610371 +0000 UTC m=+151.859278173" watchObservedRunningTime="2025-11-24 17:29:19.302770047 +0000 UTC m=+151.900437849" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.310123 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b005cb77-3bba-44a5-8446-f6fcc5053336-utilities\") pod \"redhat-operators-5wdrd\" (UID: \"b005cb77-3bba-44a5-8446-f6fcc5053336\") " pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.310218 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ea04db8-7c03-4df7-bbd5-914915bdeea0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7ea04db8-7c03-4df7-bbd5-914915bdeea0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.310367 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqn4d\" (UniqueName: \"kubernetes.io/projected/b005cb77-3bba-44a5-8446-f6fcc5053336-kube-api-access-qqn4d\") pod \"redhat-operators-5wdrd\" (UID: \"b005cb77-3bba-44a5-8446-f6fcc5053336\") " pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.310432 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b005cb77-3bba-44a5-8446-f6fcc5053336-catalog-content\") pod \"redhat-operators-5wdrd\" (UID: \"b005cb77-3bba-44a5-8446-f6fcc5053336\") " pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.310467 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ea04db8-7c03-4df7-bbd5-914915bdeea0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7ea04db8-7c03-4df7-bbd5-914915bdeea0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.310647 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b005cb77-3bba-44a5-8446-f6fcc5053336-utilities\") pod \"redhat-operators-5wdrd\" (UID: \"b005cb77-3bba-44a5-8446-f6fcc5053336\") " pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.311580 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b005cb77-3bba-44a5-8446-f6fcc5053336-catalog-content\") pod \"redhat-operators-5wdrd\" (UID: \"b005cb77-3bba-44a5-8446-f6fcc5053336\") " pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.311933 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ea04db8-7c03-4df7-bbd5-914915bdeea0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7ea04db8-7c03-4df7-bbd5-914915bdeea0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.344913 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqn4d\" (UniqueName: \"kubernetes.io/projected/b005cb77-3bba-44a5-8446-f6fcc5053336-kube-api-access-qqn4d\") pod \"redhat-operators-5wdrd\" (UID: \"b005cb77-3bba-44a5-8446-f6fcc5053336\") " pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.353772 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ea04db8-7c03-4df7-bbd5-914915bdeea0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7ea04db8-7c03-4df7-bbd5-914915bdeea0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.446333 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.513784 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bmqk8"] Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.517192 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.525988 4808 patch_prober.go:28] interesting pod/router-default-5444994796-mmrgn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:29:19 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Nov 24 17:29:19 crc kubenswrapper[4808]: [+]process-running ok Nov 24 17:29:19 crc kubenswrapper[4808]: healthz check failed Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.526185 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmrgn" podUID="3a54581b-46f9-4453-b3c0-9cb55e3410f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.548079 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bmqk8"] Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.548418 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.717126 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/234ae5d2-c9a2-470c-9137-f52babb8f9dc-utilities\") pod \"redhat-operators-bmqk8\" (UID: \"234ae5d2-c9a2-470c-9137-f52babb8f9dc\") " pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.717214 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/234ae5d2-c9a2-470c-9137-f52babb8f9dc-catalog-content\") pod \"redhat-operators-bmqk8\" (UID: \"234ae5d2-c9a2-470c-9137-f52babb8f9dc\") " pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.717266 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlf7n\" (UniqueName: \"kubernetes.io/projected/234ae5d2-c9a2-470c-9137-f52babb8f9dc-kube-api-access-zlf7n\") pod \"redhat-operators-bmqk8\" (UID: \"234ae5d2-c9a2-470c-9137-f52babb8f9dc\") " pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.818202 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/234ae5d2-c9a2-470c-9137-f52babb8f9dc-catalog-content\") pod \"redhat-operators-bmqk8\" (UID: \"234ae5d2-c9a2-470c-9137-f52babb8f9dc\") " pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.818688 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlf7n\" (UniqueName: \"kubernetes.io/projected/234ae5d2-c9a2-470c-9137-f52babb8f9dc-kube-api-access-zlf7n\") pod \"redhat-operators-bmqk8\" (UID: \"234ae5d2-c9a2-470c-9137-f52babb8f9dc\") " pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.818738 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/234ae5d2-c9a2-470c-9137-f52babb8f9dc-utilities\") pod \"redhat-operators-bmqk8\" (UID: \"234ae5d2-c9a2-470c-9137-f52babb8f9dc\") " pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.819750 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/234ae5d2-c9a2-470c-9137-f52babb8f9dc-utilities\") pod \"redhat-operators-bmqk8\" (UID: \"234ae5d2-c9a2-470c-9137-f52babb8f9dc\") " pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.819924 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/234ae5d2-c9a2-470c-9137-f52babb8f9dc-catalog-content\") pod \"redhat-operators-bmqk8\" (UID: \"234ae5d2-c9a2-470c-9137-f52babb8f9dc\") " pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.853976 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlf7n\" (UniqueName: \"kubernetes.io/projected/234ae5d2-c9a2-470c-9137-f52babb8f9dc-kube-api-access-zlf7n\") pod \"redhat-operators-bmqk8\" (UID: \"234ae5d2-c9a2-470c-9137-f52babb8f9dc\") " pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:29:19 crc kubenswrapper[4808]: I1124 17:29:19.897423 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.027889 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5wdrd"] Nov 24 17:29:20 crc kubenswrapper[4808]: W1124 17:29:20.040141 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb005cb77_3bba_44a5_8446_f6fcc5053336.slice/crio-dc3a758c84a2c6ffcec0d792ccb1cd3784c3848fa347c5aeaa071060afc79ce8 WatchSource:0}: Error finding container dc3a758c84a2c6ffcec0d792ccb1cd3784c3848fa347c5aeaa071060afc79ce8: Status 404 returned error can't find the container with id dc3a758c84a2c6ffcec0d792ccb1cd3784c3848fa347c5aeaa071060afc79ce8 Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.152916 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.152982 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.175967 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.177951 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.374770 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.376203 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.376241 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bmqk8"] Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.376279 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5wdrd" event={"ID":"b005cb77-3bba-44a5-8446-f6fcc5053336","Type":"ContainerStarted","Data":"dc3a758c84a2c6ffcec0d792ccb1cd3784c3848fa347c5aeaa071060afc79ce8"} Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.376298 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" event={"ID":"edb7cea2-f6eb-4043-91d9-c7abaef74455","Type":"ContainerStarted","Data":"f4f10a51abdab5898fea49b8f4bafb394e8a9b487712f1bf33cf4cf93623b538"} Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.376310 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" event={"ID":"edb7cea2-f6eb-4043-91d9-c7abaef74455","Type":"ContainerStarted","Data":"2586c1c9c86057fce8c557ec2d913802200bd545faab365bf2fc119b93520556"} Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.376319 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7ea04db8-7c03-4df7-bbd5-914915bdeea0","Type":"ContainerStarted","Data":"a329d323b480a0aea0224db7f1953dbb120827cdc247518e188741dff9ad9dcc"} Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.378758 4808 generic.go:334] "Generic (PLEG): container finished" podID="04baaaad-0212-49d6-86af-d498e40e8600" containerID="b0fdce55b42a3c7b0f66074ca3d7b206e30f9be44f4aaf79709780802149d99d" exitCode=0 Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.379595 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5lxz" event={"ID":"04baaaad-0212-49d6-86af-d498e40e8600","Type":"ContainerDied","Data":"b0fdce55b42a3c7b0f66074ca3d7b206e30f9be44f4aaf79709780802149d99d"} Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.386481 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7bhtk" Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.409422 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" podStartSLOduration=132.4093959 podStartE2EDuration="2m12.4093959s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:20.395077014 +0000 UTC m=+152.992744816" watchObservedRunningTime="2025-11-24 17:29:20.4093959 +0000 UTC m=+153.007063702" Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.518847 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.527193 4808 patch_prober.go:28] interesting pod/router-default-5444994796-mmrgn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:29:20 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Nov 24 17:29:20 crc kubenswrapper[4808]: [+]process-running ok Nov 24 17:29:20 crc kubenswrapper[4808]: healthz check failed Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.527292 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmrgn" podUID="3a54581b-46f9-4453-b3c0-9cb55e3410f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:29:20 crc kubenswrapper[4808]: I1124 17:29:20.806483 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xlpkq" Nov 24 17:29:21 crc kubenswrapper[4808]: I1124 17:29:21.409808 4808 generic.go:334] "Generic (PLEG): container finished" podID="b005cb77-3bba-44a5-8446-f6fcc5053336" containerID="e37a0207952e492c77dcd4b1371413b00d9a0080255c7b4dcf56ce80031d5e6f" exitCode=0 Nov 24 17:29:21 crc kubenswrapper[4808]: I1124 17:29:21.410117 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5wdrd" event={"ID":"b005cb77-3bba-44a5-8446-f6fcc5053336","Type":"ContainerDied","Data":"e37a0207952e492c77dcd4b1371413b00d9a0080255c7b4dcf56ce80031d5e6f"} Nov 24 17:29:21 crc kubenswrapper[4808]: I1124 17:29:21.427759 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7ea04db8-7c03-4df7-bbd5-914915bdeea0","Type":"ContainerStarted","Data":"1fe1e6194918d73a5278791bea1aabe401a85c02257d9e3bad1dd25e94c8e735"} Nov 24 17:29:21 crc kubenswrapper[4808]: I1124 17:29:21.441486 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmqk8" event={"ID":"234ae5d2-c9a2-470c-9137-f52babb8f9dc","Type":"ContainerStarted","Data":"dbd60cd98e2e1f08507ccbc6e249c0dacde9d58842ac44d5952e272e79fa86c4"} Nov 24 17:29:21 crc kubenswrapper[4808]: I1124 17:29:21.526164 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:21 crc kubenswrapper[4808]: I1124 17:29:21.531538 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-mmrgn" Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.286272 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.290647 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.294641 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.295539 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.300894 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.392514 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/533d67fe-95b7-49c6-941e-c035b3279626-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"533d67fe-95b7-49c6-941e-c035b3279626\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.392670 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/533d67fe-95b7-49c6-941e-c035b3279626-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"533d67fe-95b7-49c6-941e-c035b3279626\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.472531 4808 generic.go:334] "Generic (PLEG): container finished" podID="7ea04db8-7c03-4df7-bbd5-914915bdeea0" containerID="1fe1e6194918d73a5278791bea1aabe401a85c02257d9e3bad1dd25e94c8e735" exitCode=0 Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.472902 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7ea04db8-7c03-4df7-bbd5-914915bdeea0","Type":"ContainerDied","Data":"1fe1e6194918d73a5278791bea1aabe401a85c02257d9e3bad1dd25e94c8e735"} Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.477582 4808 generic.go:334] "Generic (PLEG): container finished" podID="234ae5d2-c9a2-470c-9137-f52babb8f9dc" containerID="204caa694017256631117723bd86fde7137f0bb10dbad6efa23ecd6e6893644b" exitCode=0 Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.477793 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmqk8" event={"ID":"234ae5d2-c9a2-470c-9137-f52babb8f9dc","Type":"ContainerDied","Data":"204caa694017256631117723bd86fde7137f0bb10dbad6efa23ecd6e6893644b"} Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.498849 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/533d67fe-95b7-49c6-941e-c035b3279626-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"533d67fe-95b7-49c6-941e-c035b3279626\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.498914 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/533d67fe-95b7-49c6-941e-c035b3279626-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"533d67fe-95b7-49c6-941e-c035b3279626\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.498977 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/533d67fe-95b7-49c6-941e-c035b3279626-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"533d67fe-95b7-49c6-941e-c035b3279626\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.528164 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/533d67fe-95b7-49c6-941e-c035b3279626-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"533d67fe-95b7-49c6-941e-c035b3279626\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.621335 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:29:22 crc kubenswrapper[4808]: I1124 17:29:22.998728 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 17:29:23 crc kubenswrapper[4808]: I1124 17:29:23.532679 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"533d67fe-95b7-49c6-941e-c035b3279626","Type":"ContainerStarted","Data":"4a50569839ab6694cd81576d13036122cc51fd1ccb6f751a1c3dd6846cf8064a"} Nov 24 17:29:23 crc kubenswrapper[4808]: I1124 17:29:23.919075 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:29:23 crc kubenswrapper[4808]: I1124 17:29:23.942595 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ea04db8-7c03-4df7-bbd5-914915bdeea0-kube-api-access\") pod \"7ea04db8-7c03-4df7-bbd5-914915bdeea0\" (UID: \"7ea04db8-7c03-4df7-bbd5-914915bdeea0\") " Nov 24 17:29:23 crc kubenswrapper[4808]: I1124 17:29:23.942710 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ea04db8-7c03-4df7-bbd5-914915bdeea0-kubelet-dir\") pod \"7ea04db8-7c03-4df7-bbd5-914915bdeea0\" (UID: \"7ea04db8-7c03-4df7-bbd5-914915bdeea0\") " Nov 24 17:29:23 crc kubenswrapper[4808]: I1124 17:29:23.942997 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ea04db8-7c03-4df7-bbd5-914915bdeea0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7ea04db8-7c03-4df7-bbd5-914915bdeea0" (UID: "7ea04db8-7c03-4df7-bbd5-914915bdeea0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:29:23 crc kubenswrapper[4808]: I1124 17:29:23.951209 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ea04db8-7c03-4df7-bbd5-914915bdeea0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7ea04db8-7c03-4df7-bbd5-914915bdeea0" (UID: "7ea04db8-7c03-4df7-bbd5-914915bdeea0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:29:24 crc kubenswrapper[4808]: I1124 17:29:24.044751 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ea04db8-7c03-4df7-bbd5-914915bdeea0-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 17:29:24 crc kubenswrapper[4808]: I1124 17:29:24.044797 4808 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ea04db8-7c03-4df7-bbd5-914915bdeea0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 17:29:24 crc kubenswrapper[4808]: I1124 17:29:24.558984 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:29:24 crc kubenswrapper[4808]: I1124 17:29:24.558965 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7ea04db8-7c03-4df7-bbd5-914915bdeea0","Type":"ContainerDied","Data":"a329d323b480a0aea0224db7f1953dbb120827cdc247518e188741dff9ad9dcc"} Nov 24 17:29:24 crc kubenswrapper[4808]: I1124 17:29:24.559263 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a329d323b480a0aea0224db7f1953dbb120827cdc247518e188741dff9ad9dcc" Nov 24 17:29:24 crc kubenswrapper[4808]: I1124 17:29:24.563451 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"533d67fe-95b7-49c6-941e-c035b3279626","Type":"ContainerStarted","Data":"9d74b3b54b9e65b9a4f46b3513787c47a1189d2ce9b9e7ed1458f85a96e0270a"} Nov 24 17:29:24 crc kubenswrapper[4808]: I1124 17:29:24.585691 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.5856707820000002 podStartE2EDuration="2.585670782s" podCreationTimestamp="2025-11-24 17:29:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:24.579072215 +0000 UTC m=+157.176740017" watchObservedRunningTime="2025-11-24 17:29:24.585670782 +0000 UTC m=+157.183338584" Nov 24 17:29:25 crc kubenswrapper[4808]: I1124 17:29:25.589209 4808 generic.go:334] "Generic (PLEG): container finished" podID="533d67fe-95b7-49c6-941e-c035b3279626" containerID="9d74b3b54b9e65b9a4f46b3513787c47a1189d2ce9b9e7ed1458f85a96e0270a" exitCode=0 Nov 24 17:29:25 crc kubenswrapper[4808]: I1124 17:29:25.589263 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"533d67fe-95b7-49c6-941e-c035b3279626","Type":"ContainerDied","Data":"9d74b3b54b9e65b9a4f46b3513787c47a1189d2ce9b9e7ed1458f85a96e0270a"} Nov 24 17:29:25 crc kubenswrapper[4808]: I1124 17:29:25.683729 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-2d6vb" Nov 24 17:29:28 crc kubenswrapper[4808]: I1124 17:29:28.478543 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-6xnwl" Nov 24 17:29:28 crc kubenswrapper[4808]: I1124 17:29:28.822171 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:28 crc kubenswrapper[4808]: I1124 17:29:28.827037 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:29:30 crc kubenswrapper[4808]: I1124 17:29:30.454355 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs\") pod \"network-metrics-daemon-8v2q4\" (UID: \"c26458eb-7f6c-42cb-b6c5-90de6756a994\") " pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:29:30 crc kubenswrapper[4808]: I1124 17:29:30.461753 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26458eb-7f6c-42cb-b6c5-90de6756a994-metrics-certs\") pod \"network-metrics-daemon-8v2q4\" (UID: \"c26458eb-7f6c-42cb-b6c5-90de6756a994\") " pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:29:30 crc kubenswrapper[4808]: I1124 17:29:30.692813 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8v2q4" Nov 24 17:29:36 crc kubenswrapper[4808]: I1124 17:29:36.523057 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:29:36 crc kubenswrapper[4808]: I1124 17:29:36.523432 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:29:37 crc kubenswrapper[4808]: I1124 17:29:37.896351 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:29:37 crc kubenswrapper[4808]: I1124 17:29:37.997591 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/533d67fe-95b7-49c6-941e-c035b3279626-kube-api-access\") pod \"533d67fe-95b7-49c6-941e-c035b3279626\" (UID: \"533d67fe-95b7-49c6-941e-c035b3279626\") " Nov 24 17:29:37 crc kubenswrapper[4808]: I1124 17:29:37.997734 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/533d67fe-95b7-49c6-941e-c035b3279626-kubelet-dir\") pod \"533d67fe-95b7-49c6-941e-c035b3279626\" (UID: \"533d67fe-95b7-49c6-941e-c035b3279626\") " Nov 24 17:29:37 crc kubenswrapper[4808]: I1124 17:29:37.997834 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/533d67fe-95b7-49c6-941e-c035b3279626-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "533d67fe-95b7-49c6-941e-c035b3279626" (UID: "533d67fe-95b7-49c6-941e-c035b3279626"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:29:37 crc kubenswrapper[4808]: I1124 17:29:37.998498 4808 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/533d67fe-95b7-49c6-941e-c035b3279626-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 17:29:38 crc kubenswrapper[4808]: I1124 17:29:38.004677 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/533d67fe-95b7-49c6-941e-c035b3279626-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "533d67fe-95b7-49c6-941e-c035b3279626" (UID: "533d67fe-95b7-49c6-941e-c035b3279626"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:29:38 crc kubenswrapper[4808]: I1124 17:29:38.099771 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/533d67fe-95b7-49c6-941e-c035b3279626-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 17:29:38 crc kubenswrapper[4808]: I1124 17:29:38.707587 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"533d67fe-95b7-49c6-941e-c035b3279626","Type":"ContainerDied","Data":"4a50569839ab6694cd81576d13036122cc51fd1ccb6f751a1c3dd6846cf8064a"} Nov 24 17:29:38 crc kubenswrapper[4808]: I1124 17:29:38.708378 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a50569839ab6694cd81576d13036122cc51fd1ccb6f751a1c3dd6846cf8064a" Nov 24 17:29:38 crc kubenswrapper[4808]: I1124 17:29:38.707647 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:29:38 crc kubenswrapper[4808]: I1124 17:29:38.885082 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:29:48 crc kubenswrapper[4808]: E1124 17:29:48.887680 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 24 17:29:48 crc kubenswrapper[4808]: E1124 17:29:48.889317 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-thjx9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-4pwnf_openshift-marketplace(7a0cb760-04cf-4658-b4ff-1da560789800): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 17:29:48 crc kubenswrapper[4808]: E1124 17:29:48.891408 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-4pwnf" podUID="7a0cb760-04cf-4658-b4ff-1da560789800" Nov 24 17:29:50 crc kubenswrapper[4808]: E1124 17:29:50.184791 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-4pwnf" podUID="7a0cb760-04cf-4658-b4ff-1da560789800" Nov 24 17:29:50 crc kubenswrapper[4808]: I1124 17:29:50.584765 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wsfw6" Nov 24 17:29:53 crc kubenswrapper[4808]: E1124 17:29:53.386512 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 24 17:29:53 crc kubenswrapper[4808]: E1124 17:29:53.387746 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zlf7n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-bmqk8_openshift-marketplace(234ae5d2-c9a2-470c-9137-f52babb8f9dc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 17:29:53 crc kubenswrapper[4808]: E1124 17:29:53.390487 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-bmqk8" podUID="234ae5d2-c9a2-470c-9137-f52babb8f9dc" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.439188 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-bmqk8" podUID="234ae5d2-c9a2-470c-9137-f52babb8f9dc" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.545351 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.545480 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2m966,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-v5lxz_openshift-marketplace(04baaaad-0212-49d6-86af-d498e40e8600): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.546630 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-v5lxz" podUID="04baaaad-0212-49d6-86af-d498e40e8600" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.558623 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.558768 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v2r6w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-96cn9_openshift-marketplace(cde9313c-9316-40dc-8427-a8683ce67b5b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.559973 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-96cn9" podUID="cde9313c-9316-40dc-8427-a8683ce67b5b" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.591951 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.592127 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w4g78,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-fwnlt_openshift-marketplace(f11d89a2-be43-47fd-ae09-57fcd7df9820): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.593309 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-fwnlt" podUID="f11d89a2-be43-47fd-ae09-57fcd7df9820" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.616278 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.616460 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qqn4d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-5wdrd_openshift-marketplace(b005cb77-3bba-44a5-8446-f6fcc5053336): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.617688 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-5wdrd" podUID="b005cb77-3bba-44a5-8446-f6fcc5053336" Nov 24 17:29:54 crc kubenswrapper[4808]: I1124 17:29:54.670638 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-8v2q4"] Nov 24 17:29:54 crc kubenswrapper[4808]: W1124 17:29:54.695629 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc26458eb_7f6c_42cb_b6c5_90de6756a994.slice/crio-bb4de3fe4bcfe54c29399e0e1c3c5508a7c47fc9bf5eb31db1cfb4046f3004d2 WatchSource:0}: Error finding container bb4de3fe4bcfe54c29399e0e1c3c5508a7c47fc9bf5eb31db1cfb4046f3004d2: Status 404 returned error can't find the container with id bb4de3fe4bcfe54c29399e0e1c3c5508a7c47fc9bf5eb31db1cfb4046f3004d2 Nov 24 17:29:54 crc kubenswrapper[4808]: I1124 17:29:54.820175 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j9jpf" event={"ID":"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3","Type":"ContainerStarted","Data":"a39cb9c4cb78f574b63fe2aae5e9f1ea980de5f1590395f95a0b20f68d8becb8"} Nov 24 17:29:54 crc kubenswrapper[4808]: I1124 17:29:54.821750 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" event={"ID":"c26458eb-7f6c-42cb-b6c5-90de6756a994","Type":"ContainerStarted","Data":"bb4de3fe4bcfe54c29399e0e1c3c5508a7c47fc9bf5eb31db1cfb4046f3004d2"} Nov 24 17:29:54 crc kubenswrapper[4808]: I1124 17:29:54.824327 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlmvz" event={"ID":"a3f2246d-a6bb-463f-9a38-2d7144079eb6","Type":"ContainerStarted","Data":"06d164f15f4ffda32fc38e8dd0fc8362cd4ad5865f18862e868ad682cb37fbff"} Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.825726 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-v5lxz" podUID="04baaaad-0212-49d6-86af-d498e40e8600" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.826811 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-96cn9" podUID="cde9313c-9316-40dc-8427-a8683ce67b5b" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.855787 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-5wdrd" podUID="b005cb77-3bba-44a5-8446-f6fcc5053336" Nov 24 17:29:54 crc kubenswrapper[4808]: E1124 17:29:54.866895 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-fwnlt" podUID="f11d89a2-be43-47fd-ae09-57fcd7df9820" Nov 24 17:29:55 crc kubenswrapper[4808]: I1124 17:29:55.832416 4808 generic.go:334] "Generic (PLEG): container finished" podID="b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" containerID="a39cb9c4cb78f574b63fe2aae5e9f1ea980de5f1590395f95a0b20f68d8becb8" exitCode=0 Nov 24 17:29:55 crc kubenswrapper[4808]: I1124 17:29:55.832501 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j9jpf" event={"ID":"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3","Type":"ContainerDied","Data":"a39cb9c4cb78f574b63fe2aae5e9f1ea980de5f1590395f95a0b20f68d8becb8"} Nov 24 17:29:55 crc kubenswrapper[4808]: I1124 17:29:55.837248 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" event={"ID":"c26458eb-7f6c-42cb-b6c5-90de6756a994","Type":"ContainerStarted","Data":"61ab97fef5310969a9c75934d2009032e4c361908758cf8a0274b4696c90dd1b"} Nov 24 17:29:55 crc kubenswrapper[4808]: I1124 17:29:55.837304 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8v2q4" event={"ID":"c26458eb-7f6c-42cb-b6c5-90de6756a994","Type":"ContainerStarted","Data":"0ab0294632e8dab248e76e65227de5b28431bb4542e0d8da3176a8b99ca6141e"} Nov 24 17:29:55 crc kubenswrapper[4808]: I1124 17:29:55.841961 4808 generic.go:334] "Generic (PLEG): container finished" podID="a3f2246d-a6bb-463f-9a38-2d7144079eb6" containerID="06d164f15f4ffda32fc38e8dd0fc8362cd4ad5865f18862e868ad682cb37fbff" exitCode=0 Nov 24 17:29:55 crc kubenswrapper[4808]: I1124 17:29:55.842049 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlmvz" event={"ID":"a3f2246d-a6bb-463f-9a38-2d7144079eb6","Type":"ContainerDied","Data":"06d164f15f4ffda32fc38e8dd0fc8362cd4ad5865f18862e868ad682cb37fbff"} Nov 24 17:29:55 crc kubenswrapper[4808]: I1124 17:29:55.898222 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-8v2q4" podStartSLOduration=167.898204324 podStartE2EDuration="2m47.898204324s" podCreationTimestamp="2025-11-24 17:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:29:55.872932091 +0000 UTC m=+188.470599913" watchObservedRunningTime="2025-11-24 17:29:55.898204324 +0000 UTC m=+188.495872126" Nov 24 17:29:56 crc kubenswrapper[4808]: I1124 17:29:56.588857 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:29:56 crc kubenswrapper[4808]: I1124 17:29:56.849888 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlmvz" event={"ID":"a3f2246d-a6bb-463f-9a38-2d7144079eb6","Type":"ContainerStarted","Data":"c2e447480116b74cbcb702d927fd337acfa8eb284076997d3fe05668eb7853ac"} Nov 24 17:29:56 crc kubenswrapper[4808]: I1124 17:29:56.852150 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j9jpf" event={"ID":"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3","Type":"ContainerStarted","Data":"6b054a87da2e010285f250e1bb1bb578abfed98903b862039dad79254cabb562"} Nov 24 17:29:56 crc kubenswrapper[4808]: I1124 17:29:56.871789 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mlmvz" podStartSLOduration=2.70243628 podStartE2EDuration="40.871765485s" podCreationTimestamp="2025-11-24 17:29:16 +0000 UTC" firstStartedPulling="2025-11-24 17:29:18.112848903 +0000 UTC m=+150.710516705" lastFinishedPulling="2025-11-24 17:29:56.282178108 +0000 UTC m=+188.879845910" observedRunningTime="2025-11-24 17:29:56.870644922 +0000 UTC m=+189.468312724" watchObservedRunningTime="2025-11-24 17:29:56.871765485 +0000 UTC m=+189.469433287" Nov 24 17:29:56 crc kubenswrapper[4808]: I1124 17:29:56.892304 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j9jpf" podStartSLOduration=3.95235801 podStartE2EDuration="40.892277816s" podCreationTimestamp="2025-11-24 17:29:16 +0000 UTC" firstStartedPulling="2025-11-24 17:29:19.280853374 +0000 UTC m=+151.878521176" lastFinishedPulling="2025-11-24 17:29:56.22077318 +0000 UTC m=+188.818440982" observedRunningTime="2025-11-24 17:29:56.891917495 +0000 UTC m=+189.489585297" watchObservedRunningTime="2025-11-24 17:29:56.892277816 +0000 UTC m=+189.489945618" Nov 24 17:29:56 crc kubenswrapper[4808]: I1124 17:29:56.939154 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:29:56 crc kubenswrapper[4808]: I1124 17:29:56.939211 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:29:58 crc kubenswrapper[4808]: I1124 17:29:58.080208 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-j9jpf" podUID="b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" containerName="registry-server" probeResult="failure" output=< Nov 24 17:29:58 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Nov 24 17:29:58 crc kubenswrapper[4808]: > Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.143303 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz"] Nov 24 17:30:00 crc kubenswrapper[4808]: E1124 17:30:00.143728 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="533d67fe-95b7-49c6-941e-c035b3279626" containerName="pruner" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.143754 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="533d67fe-95b7-49c6-941e-c035b3279626" containerName="pruner" Nov 24 17:30:00 crc kubenswrapper[4808]: E1124 17:30:00.143789 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea04db8-7c03-4df7-bbd5-914915bdeea0" containerName="pruner" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.143798 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea04db8-7c03-4df7-bbd5-914915bdeea0" containerName="pruner" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.143937 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="533d67fe-95b7-49c6-941e-c035b3279626" containerName="pruner" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.143958 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ea04db8-7c03-4df7-bbd5-914915bdeea0" containerName="pruner" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.144689 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.147824 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.148281 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.148356 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz"] Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.160414 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-config-volume\") pod \"collect-profiles-29400090-d86cz\" (UID: \"fee761ab-3ef8-4ee5-9bea-59395abe3fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.160696 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmk2j\" (UniqueName: \"kubernetes.io/projected/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-kube-api-access-vmk2j\") pod \"collect-profiles-29400090-d86cz\" (UID: \"fee761ab-3ef8-4ee5-9bea-59395abe3fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.160785 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-secret-volume\") pod \"collect-profiles-29400090-d86cz\" (UID: \"fee761ab-3ef8-4ee5-9bea-59395abe3fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.275275 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-config-volume\") pod \"collect-profiles-29400090-d86cz\" (UID: \"fee761ab-3ef8-4ee5-9bea-59395abe3fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.276040 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmk2j\" (UniqueName: \"kubernetes.io/projected/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-kube-api-access-vmk2j\") pod \"collect-profiles-29400090-d86cz\" (UID: \"fee761ab-3ef8-4ee5-9bea-59395abe3fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.276200 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-secret-volume\") pod \"collect-profiles-29400090-d86cz\" (UID: \"fee761ab-3ef8-4ee5-9bea-59395abe3fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.277746 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-config-volume\") pod \"collect-profiles-29400090-d86cz\" (UID: \"fee761ab-3ef8-4ee5-9bea-59395abe3fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.285131 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-secret-volume\") pod \"collect-profiles-29400090-d86cz\" (UID: \"fee761ab-3ef8-4ee5-9bea-59395abe3fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.296457 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmk2j\" (UniqueName: \"kubernetes.io/projected/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-kube-api-access-vmk2j\") pod \"collect-profiles-29400090-d86cz\" (UID: \"fee761ab-3ef8-4ee5-9bea-59395abe3fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.470226 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.661778 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz"] Nov 24 17:30:00 crc kubenswrapper[4808]: W1124 17:30:00.675458 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfee761ab_3ef8_4ee5_9bea_59395abe3fbc.slice/crio-970883a7211d3981f3a856522beadadf3b33376fdc030be662893a1f7d645b6f WatchSource:0}: Error finding container 970883a7211d3981f3a856522beadadf3b33376fdc030be662893a1f7d645b6f: Status 404 returned error can't find the container with id 970883a7211d3981f3a856522beadadf3b33376fdc030be662893a1f7d645b6f Nov 24 17:30:00 crc kubenswrapper[4808]: I1124 17:30:00.892338 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" event={"ID":"fee761ab-3ef8-4ee5-9bea-59395abe3fbc","Type":"ContainerStarted","Data":"970883a7211d3981f3a856522beadadf3b33376fdc030be662893a1f7d645b6f"} Nov 24 17:30:01 crc kubenswrapper[4808]: I1124 17:30:01.900643 4808 generic.go:334] "Generic (PLEG): container finished" podID="fee761ab-3ef8-4ee5-9bea-59395abe3fbc" containerID="207281a72bfd1685b869102173cb9465159e31d100b42e88c9039e7273c86b88" exitCode=0 Nov 24 17:30:01 crc kubenswrapper[4808]: I1124 17:30:01.900731 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" event={"ID":"fee761ab-3ef8-4ee5-9bea-59395abe3fbc","Type":"ContainerDied","Data":"207281a72bfd1685b869102173cb9465159e31d100b42e88c9039e7273c86b88"} Nov 24 17:30:03 crc kubenswrapper[4808]: I1124 17:30:03.119180 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" Nov 24 17:30:03 crc kubenswrapper[4808]: I1124 17:30:03.223324 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-config-volume\") pod \"fee761ab-3ef8-4ee5-9bea-59395abe3fbc\" (UID: \"fee761ab-3ef8-4ee5-9bea-59395abe3fbc\") " Nov 24 17:30:03 crc kubenswrapper[4808]: I1124 17:30:03.223399 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-secret-volume\") pod \"fee761ab-3ef8-4ee5-9bea-59395abe3fbc\" (UID: \"fee761ab-3ef8-4ee5-9bea-59395abe3fbc\") " Nov 24 17:30:03 crc kubenswrapper[4808]: I1124 17:30:03.223544 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmk2j\" (UniqueName: \"kubernetes.io/projected/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-kube-api-access-vmk2j\") pod \"fee761ab-3ef8-4ee5-9bea-59395abe3fbc\" (UID: \"fee761ab-3ef8-4ee5-9bea-59395abe3fbc\") " Nov 24 17:30:03 crc kubenswrapper[4808]: I1124 17:30:03.224626 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-config-volume" (OuterVolumeSpecName: "config-volume") pod "fee761ab-3ef8-4ee5-9bea-59395abe3fbc" (UID: "fee761ab-3ef8-4ee5-9bea-59395abe3fbc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:30:03 crc kubenswrapper[4808]: I1124 17:30:03.231215 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-kube-api-access-vmk2j" (OuterVolumeSpecName: "kube-api-access-vmk2j") pod "fee761ab-3ef8-4ee5-9bea-59395abe3fbc" (UID: "fee761ab-3ef8-4ee5-9bea-59395abe3fbc"). InnerVolumeSpecName "kube-api-access-vmk2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:30:03 crc kubenswrapper[4808]: I1124 17:30:03.231384 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fee761ab-3ef8-4ee5-9bea-59395abe3fbc" (UID: "fee761ab-3ef8-4ee5-9bea-59395abe3fbc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:30:03 crc kubenswrapper[4808]: I1124 17:30:03.324989 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmk2j\" (UniqueName: \"kubernetes.io/projected/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-kube-api-access-vmk2j\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:03 crc kubenswrapper[4808]: I1124 17:30:03.325052 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:03 crc kubenswrapper[4808]: I1124 17:30:03.325091 4808 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fee761ab-3ef8-4ee5-9bea-59395abe3fbc-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:03 crc kubenswrapper[4808]: I1124 17:30:03.913758 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" event={"ID":"fee761ab-3ef8-4ee5-9bea-59395abe3fbc","Type":"ContainerDied","Data":"970883a7211d3981f3a856522beadadf3b33376fdc030be662893a1f7d645b6f"} Nov 24 17:30:03 crc kubenswrapper[4808]: I1124 17:30:03.913866 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="970883a7211d3981f3a856522beadadf3b33376fdc030be662893a1f7d645b6f" Nov 24 17:30:03 crc kubenswrapper[4808]: I1124 17:30:03.913903 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz" Nov 24 17:30:06 crc kubenswrapper[4808]: I1124 17:30:06.436671 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:30:06 crc kubenswrapper[4808]: I1124 17:30:06.437226 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:30:06 crc kubenswrapper[4808]: I1124 17:30:06.502820 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:30:06 crc kubenswrapper[4808]: I1124 17:30:06.522837 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:30:06 crc kubenswrapper[4808]: I1124 17:30:06.522915 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:30:06 crc kubenswrapper[4808]: I1124 17:30:06.979410 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:30:06 crc kubenswrapper[4808]: I1124 17:30:06.985973 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:30:07 crc kubenswrapper[4808]: I1124 17:30:07.042249 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:30:07 crc kubenswrapper[4808]: I1124 17:30:07.934759 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j9jpf"] Nov 24 17:30:07 crc kubenswrapper[4808]: I1124 17:30:07.937439 4808 generic.go:334] "Generic (PLEG): container finished" podID="7a0cb760-04cf-4658-b4ff-1da560789800" containerID="f7b59228cee8bd2dd1f5862664f50315ac05d1be0d607f8917e40ae1000b115b" exitCode=0 Nov 24 17:30:07 crc kubenswrapper[4808]: I1124 17:30:07.937524 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4pwnf" event={"ID":"7a0cb760-04cf-4658-b4ff-1da560789800","Type":"ContainerDied","Data":"f7b59228cee8bd2dd1f5862664f50315ac05d1be0d607f8917e40ae1000b115b"} Nov 24 17:30:08 crc kubenswrapper[4808]: I1124 17:30:08.946925 4808 generic.go:334] "Generic (PLEG): container finished" podID="04baaaad-0212-49d6-86af-d498e40e8600" containerID="ff96358102311368af9e4d45a5f370a6c053cd6bc57e28ec9b3b756beb134df6" exitCode=0 Nov 24 17:30:08 crc kubenswrapper[4808]: I1124 17:30:08.946996 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5lxz" event={"ID":"04baaaad-0212-49d6-86af-d498e40e8600","Type":"ContainerDied","Data":"ff96358102311368af9e4d45a5f370a6c053cd6bc57e28ec9b3b756beb134df6"} Nov 24 17:30:08 crc kubenswrapper[4808]: I1124 17:30:08.956396 4808 generic.go:334] "Generic (PLEG): container finished" podID="234ae5d2-c9a2-470c-9137-f52babb8f9dc" containerID="08d18bc0d536a68b8f5a0e38ddd17fe6d4a2f7bc8f778c57aa777ceda7cbe6a5" exitCode=0 Nov 24 17:30:08 crc kubenswrapper[4808]: I1124 17:30:08.956497 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmqk8" event={"ID":"234ae5d2-c9a2-470c-9137-f52babb8f9dc","Type":"ContainerDied","Data":"08d18bc0d536a68b8f5a0e38ddd17fe6d4a2f7bc8f778c57aa777ceda7cbe6a5"} Nov 24 17:30:08 crc kubenswrapper[4808]: I1124 17:30:08.961297 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4pwnf" event={"ID":"7a0cb760-04cf-4658-b4ff-1da560789800","Type":"ContainerStarted","Data":"fe3751e91bad9cee48532b170eb25d59873a1bb344d42adf0031d6cfb4ae765c"} Nov 24 17:30:08 crc kubenswrapper[4808]: I1124 17:30:08.961371 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j9jpf" podUID="b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" containerName="registry-server" containerID="cri-o://6b054a87da2e010285f250e1bb1bb578abfed98903b862039dad79254cabb562" gracePeriod=2 Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.015381 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4pwnf" podStartSLOduration=3.747131723 podStartE2EDuration="53.015364465s" podCreationTimestamp="2025-11-24 17:29:16 +0000 UTC" firstStartedPulling="2025-11-24 17:29:19.198188923 +0000 UTC m=+151.795856725" lastFinishedPulling="2025-11-24 17:30:08.466421665 +0000 UTC m=+201.064089467" observedRunningTime="2025-11-24 17:30:09.011640013 +0000 UTC m=+201.609307815" watchObservedRunningTime="2025-11-24 17:30:09.015364465 +0000 UTC m=+201.613032267" Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.335262 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.454571 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp88c\" (UniqueName: \"kubernetes.io/projected/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-kube-api-access-wp88c\") pod \"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3\" (UID: \"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3\") " Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.454659 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-utilities\") pod \"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3\" (UID: \"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3\") " Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.454782 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-catalog-content\") pod \"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3\" (UID: \"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3\") " Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.456687 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-utilities" (OuterVolumeSpecName: "utilities") pod "b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" (UID: "b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.468003 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-kube-api-access-wp88c" (OuterVolumeSpecName: "kube-api-access-wp88c") pod "b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" (UID: "b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3"). InnerVolumeSpecName "kube-api-access-wp88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.523398 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" (UID: "b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.556278 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp88c\" (UniqueName: \"kubernetes.io/projected/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-kube-api-access-wp88c\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.556323 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.556336 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.973616 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmqk8" event={"ID":"234ae5d2-c9a2-470c-9137-f52babb8f9dc","Type":"ContainerStarted","Data":"49d65bd2faf950e4b9d8eabe9ec24da5c75c0f0233c1dbbd39d08e003afdb061"} Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.976201 4808 generic.go:334] "Generic (PLEG): container finished" podID="b005cb77-3bba-44a5-8446-f6fcc5053336" containerID="d243094397bae39dc1bcdc98fafeadb91f71740003cda8386cd069d03bc0b92c" exitCode=0 Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.976312 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5wdrd" event={"ID":"b005cb77-3bba-44a5-8446-f6fcc5053336","Type":"ContainerDied","Data":"d243094397bae39dc1bcdc98fafeadb91f71740003cda8386cd069d03bc0b92c"} Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.978732 4808 generic.go:334] "Generic (PLEG): container finished" podID="b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" containerID="6b054a87da2e010285f250e1bb1bb578abfed98903b862039dad79254cabb562" exitCode=0 Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.978775 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j9jpf" event={"ID":"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3","Type":"ContainerDied","Data":"6b054a87da2e010285f250e1bb1bb578abfed98903b862039dad79254cabb562"} Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.978822 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j9jpf" event={"ID":"b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3","Type":"ContainerDied","Data":"e6780a3abcfb50bb0392c0c4786e833d64d6a91750ec1f9be87573b4b1853014"} Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.978843 4808 scope.go:117] "RemoveContainer" containerID="6b054a87da2e010285f250e1bb1bb578abfed98903b862039dad79254cabb562" Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.978869 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j9jpf" Nov 24 17:30:09 crc kubenswrapper[4808]: I1124 17:30:09.984009 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5lxz" event={"ID":"04baaaad-0212-49d6-86af-d498e40e8600","Type":"ContainerStarted","Data":"722e267ea35fb19265ac1ce302ba43e6d57d2d65cdcbb5b8c2788786c5410989"} Nov 24 17:30:10 crc kubenswrapper[4808]: I1124 17:30:10.009467 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bmqk8" podStartSLOduration=4.079357822 podStartE2EDuration="51.009435843s" podCreationTimestamp="2025-11-24 17:29:19 +0000 UTC" firstStartedPulling="2025-11-24 17:29:22.48239634 +0000 UTC m=+155.080064142" lastFinishedPulling="2025-11-24 17:30:09.412474361 +0000 UTC m=+202.010142163" observedRunningTime="2025-11-24 17:30:10.008785533 +0000 UTC m=+202.606453335" watchObservedRunningTime="2025-11-24 17:30:10.009435843 +0000 UTC m=+202.607103635" Nov 24 17:30:10 crc kubenswrapper[4808]: I1124 17:30:10.011129 4808 scope.go:117] "RemoveContainer" containerID="a39cb9c4cb78f574b63fe2aae5e9f1ea980de5f1590395f95a0b20f68d8becb8" Nov 24 17:30:10 crc kubenswrapper[4808]: I1124 17:30:10.041253 4808 scope.go:117] "RemoveContainer" containerID="af9fedb99208e3b636087707a803b50e5a7e32756663b13291a7cc892986231d" Nov 24 17:30:10 crc kubenswrapper[4808]: I1124 17:30:10.042098 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j9jpf"] Nov 24 17:30:10 crc kubenswrapper[4808]: I1124 17:30:10.058365 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j9jpf"] Nov 24 17:30:10 crc kubenswrapper[4808]: I1124 17:30:10.059950 4808 scope.go:117] "RemoveContainer" containerID="6b054a87da2e010285f250e1bb1bb578abfed98903b862039dad79254cabb562" Nov 24 17:30:10 crc kubenswrapper[4808]: E1124 17:30:10.060695 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b054a87da2e010285f250e1bb1bb578abfed98903b862039dad79254cabb562\": container with ID starting with 6b054a87da2e010285f250e1bb1bb578abfed98903b862039dad79254cabb562 not found: ID does not exist" containerID="6b054a87da2e010285f250e1bb1bb578abfed98903b862039dad79254cabb562" Nov 24 17:30:10 crc kubenswrapper[4808]: I1124 17:30:10.060755 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b054a87da2e010285f250e1bb1bb578abfed98903b862039dad79254cabb562"} err="failed to get container status \"6b054a87da2e010285f250e1bb1bb578abfed98903b862039dad79254cabb562\": rpc error: code = NotFound desc = could not find container \"6b054a87da2e010285f250e1bb1bb578abfed98903b862039dad79254cabb562\": container with ID starting with 6b054a87da2e010285f250e1bb1bb578abfed98903b862039dad79254cabb562 not found: ID does not exist" Nov 24 17:30:10 crc kubenswrapper[4808]: I1124 17:30:10.060836 4808 scope.go:117] "RemoveContainer" containerID="a39cb9c4cb78f574b63fe2aae5e9f1ea980de5f1590395f95a0b20f68d8becb8" Nov 24 17:30:10 crc kubenswrapper[4808]: E1124 17:30:10.061578 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a39cb9c4cb78f574b63fe2aae5e9f1ea980de5f1590395f95a0b20f68d8becb8\": container with ID starting with a39cb9c4cb78f574b63fe2aae5e9f1ea980de5f1590395f95a0b20f68d8becb8 not found: ID does not exist" containerID="a39cb9c4cb78f574b63fe2aae5e9f1ea980de5f1590395f95a0b20f68d8becb8" Nov 24 17:30:10 crc kubenswrapper[4808]: I1124 17:30:10.061667 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a39cb9c4cb78f574b63fe2aae5e9f1ea980de5f1590395f95a0b20f68d8becb8"} err="failed to get container status \"a39cb9c4cb78f574b63fe2aae5e9f1ea980de5f1590395f95a0b20f68d8becb8\": rpc error: code = NotFound desc = could not find container \"a39cb9c4cb78f574b63fe2aae5e9f1ea980de5f1590395f95a0b20f68d8becb8\": container with ID starting with a39cb9c4cb78f574b63fe2aae5e9f1ea980de5f1590395f95a0b20f68d8becb8 not found: ID does not exist" Nov 24 17:30:10 crc kubenswrapper[4808]: I1124 17:30:10.061754 4808 scope.go:117] "RemoveContainer" containerID="af9fedb99208e3b636087707a803b50e5a7e32756663b13291a7cc892986231d" Nov 24 17:30:10 crc kubenswrapper[4808]: E1124 17:30:10.062219 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af9fedb99208e3b636087707a803b50e5a7e32756663b13291a7cc892986231d\": container with ID starting with af9fedb99208e3b636087707a803b50e5a7e32756663b13291a7cc892986231d not found: ID does not exist" containerID="af9fedb99208e3b636087707a803b50e5a7e32756663b13291a7cc892986231d" Nov 24 17:30:10 crc kubenswrapper[4808]: I1124 17:30:10.062265 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af9fedb99208e3b636087707a803b50e5a7e32756663b13291a7cc892986231d"} err="failed to get container status \"af9fedb99208e3b636087707a803b50e5a7e32756663b13291a7cc892986231d\": rpc error: code = NotFound desc = could not find container \"af9fedb99208e3b636087707a803b50e5a7e32756663b13291a7cc892986231d\": container with ID starting with af9fedb99208e3b636087707a803b50e5a7e32756663b13291a7cc892986231d not found: ID does not exist" Nov 24 17:30:10 crc kubenswrapper[4808]: I1124 17:30:10.081703 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v5lxz" podStartSLOduration=3.134175569 podStartE2EDuration="52.081669429s" podCreationTimestamp="2025-11-24 17:29:18 +0000 UTC" firstStartedPulling="2025-11-24 17:29:20.388219729 +0000 UTC m=+152.985887531" lastFinishedPulling="2025-11-24 17:30:09.335713589 +0000 UTC m=+201.933381391" observedRunningTime="2025-11-24 17:30:10.064965831 +0000 UTC m=+202.662633633" watchObservedRunningTime="2025-11-24 17:30:10.081669429 +0000 UTC m=+202.679337231" Nov 24 17:30:10 crc kubenswrapper[4808]: I1124 17:30:10.354541 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" path="/var/lib/kubelet/pods/b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3/volumes" Nov 24 17:30:10 crc kubenswrapper[4808]: E1124 17:30:10.545523 4808 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcde9313c_9316_40dc_8427_a8683ce67b5b.slice/crio-conmon-a20f267a55e4331b4c5d56204677052a620e963ed21bd48f5c83bd080d6e5e83.scope\": RecentStats: unable to find data in memory cache]" Nov 24 17:30:10 crc kubenswrapper[4808]: I1124 17:30:10.993054 4808 generic.go:334] "Generic (PLEG): container finished" podID="cde9313c-9316-40dc-8427-a8683ce67b5b" containerID="a20f267a55e4331b4c5d56204677052a620e963ed21bd48f5c83bd080d6e5e83" exitCode=0 Nov 24 17:30:10 crc kubenswrapper[4808]: I1124 17:30:10.994082 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96cn9" event={"ID":"cde9313c-9316-40dc-8427-a8683ce67b5b","Type":"ContainerDied","Data":"a20f267a55e4331b4c5d56204677052a620e963ed21bd48f5c83bd080d6e5e83"} Nov 24 17:30:16 crc kubenswrapper[4808]: I1124 17:30:16.022342 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fwnlt" event={"ID":"f11d89a2-be43-47fd-ae09-57fcd7df9820","Type":"ContainerStarted","Data":"bbad2b38eeb4ef59ee7db3e9900cd9e50c77515e0f1b3ffd815c8b4c7cdefd99"} Nov 24 17:30:16 crc kubenswrapper[4808]: I1124 17:30:16.025162 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5wdrd" event={"ID":"b005cb77-3bba-44a5-8446-f6fcc5053336","Type":"ContainerStarted","Data":"2d7a3cb59965a53140d64612d3ab007f76ea1382eeef72510e0d469dee1c791a"} Nov 24 17:30:16 crc kubenswrapper[4808]: I1124 17:30:16.027824 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96cn9" event={"ID":"cde9313c-9316-40dc-8427-a8683ce67b5b","Type":"ContainerStarted","Data":"69001147dd08c93f8adef7b397ecfe5c5cdd8992418d86e2a9624b691cdeee11"} Nov 24 17:30:16 crc kubenswrapper[4808]: I1124 17:30:16.075226 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5wdrd" podStartSLOduration=6.13507106 podStartE2EDuration="57.075197912s" podCreationTimestamp="2025-11-24 17:29:19 +0000 UTC" firstStartedPulling="2025-11-24 17:29:21.412129859 +0000 UTC m=+154.009797661" lastFinishedPulling="2025-11-24 17:30:12.352256711 +0000 UTC m=+204.949924513" observedRunningTime="2025-11-24 17:30:16.071176122 +0000 UTC m=+208.668843944" watchObservedRunningTime="2025-11-24 17:30:16.075197912 +0000 UTC m=+208.672865714" Nov 24 17:30:17 crc kubenswrapper[4808]: I1124 17:30:17.036535 4808 generic.go:334] "Generic (PLEG): container finished" podID="f11d89a2-be43-47fd-ae09-57fcd7df9820" containerID="bbad2b38eeb4ef59ee7db3e9900cd9e50c77515e0f1b3ffd815c8b4c7cdefd99" exitCode=0 Nov 24 17:30:17 crc kubenswrapper[4808]: I1124 17:30:17.036646 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fwnlt" event={"ID":"f11d89a2-be43-47fd-ae09-57fcd7df9820","Type":"ContainerDied","Data":"bbad2b38eeb4ef59ee7db3e9900cd9e50c77515e0f1b3ffd815c8b4c7cdefd99"} Nov 24 17:30:17 crc kubenswrapper[4808]: I1124 17:30:17.069820 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-96cn9" podStartSLOduration=4.957773005 podStartE2EDuration="1m0.069794776s" podCreationTimestamp="2025-11-24 17:29:17 +0000 UTC" firstStartedPulling="2025-11-24 17:29:19.286595235 +0000 UTC m=+151.884263037" lastFinishedPulling="2025-11-24 17:30:14.398617006 +0000 UTC m=+206.996284808" observedRunningTime="2025-11-24 17:30:17.067874759 +0000 UTC m=+209.665542561" watchObservedRunningTime="2025-11-24 17:30:17.069794776 +0000 UTC m=+209.667462578" Nov 24 17:30:17 crc kubenswrapper[4808]: I1124 17:30:17.155743 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:30:17 crc kubenswrapper[4808]: I1124 17:30:17.155792 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:30:17 crc kubenswrapper[4808]: I1124 17:30:17.199828 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:30:18 crc kubenswrapper[4808]: I1124 17:30:18.044350 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fwnlt" event={"ID":"f11d89a2-be43-47fd-ae09-57fcd7df9820","Type":"ContainerStarted","Data":"232dc4691cf01a2e914d3b89c3027a795a276f026a1fd2558510f1e52e155a22"} Nov 24 17:30:18 crc kubenswrapper[4808]: I1124 17:30:18.062435 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fwnlt" podStartSLOduration=3.6631655739999998 podStartE2EDuration="1m3.062407982s" podCreationTimestamp="2025-11-24 17:29:15 +0000 UTC" firstStartedPulling="2025-11-24 17:29:18.103590477 +0000 UTC m=+150.701258269" lastFinishedPulling="2025-11-24 17:30:17.502832845 +0000 UTC m=+210.100500677" observedRunningTime="2025-11-24 17:30:18.062310619 +0000 UTC m=+210.659978431" watchObservedRunningTime="2025-11-24 17:30:18.062407982 +0000 UTC m=+210.660075774" Nov 24 17:30:18 crc kubenswrapper[4808]: I1124 17:30:18.088996 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:30:18 crc kubenswrapper[4808]: I1124 17:30:18.238324 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:30:18 crc kubenswrapper[4808]: I1124 17:30:18.238402 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:30:18 crc kubenswrapper[4808]: I1124 17:30:18.277837 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:30:18 crc kubenswrapper[4808]: I1124 17:30:18.684913 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:30:18 crc kubenswrapper[4808]: I1124 17:30:18.684994 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:30:18 crc kubenswrapper[4808]: I1124 17:30:18.729632 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:30:19 crc kubenswrapper[4808]: I1124 17:30:19.086562 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:30:19 crc kubenswrapper[4808]: I1124 17:30:19.334735 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4pwnf"] Nov 24 17:30:19 crc kubenswrapper[4808]: I1124 17:30:19.449052 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:30:19 crc kubenswrapper[4808]: I1124 17:30:19.449153 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:30:19 crc kubenswrapper[4808]: I1124 17:30:19.898766 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:30:19 crc kubenswrapper[4808]: I1124 17:30:19.898830 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:30:19 crc kubenswrapper[4808]: I1124 17:30:19.942681 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:30:20 crc kubenswrapper[4808]: I1124 17:30:20.056188 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4pwnf" podUID="7a0cb760-04cf-4658-b4ff-1da560789800" containerName="registry-server" containerID="cri-o://fe3751e91bad9cee48532b170eb25d59873a1bb344d42adf0031d6cfb4ae765c" gracePeriod=2 Nov 24 17:30:20 crc kubenswrapper[4808]: I1124 17:30:20.105420 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:30:20 crc kubenswrapper[4808]: I1124 17:30:20.397820 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:30:20 crc kubenswrapper[4808]: I1124 17:30:20.490824 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5wdrd" podUID="b005cb77-3bba-44a5-8446-f6fcc5053336" containerName="registry-server" probeResult="failure" output=< Nov 24 17:30:20 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Nov 24 17:30:20 crc kubenswrapper[4808]: > Nov 24 17:30:20 crc kubenswrapper[4808]: I1124 17:30:20.522853 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a0cb760-04cf-4658-b4ff-1da560789800-utilities\") pod \"7a0cb760-04cf-4658-b4ff-1da560789800\" (UID: \"7a0cb760-04cf-4658-b4ff-1da560789800\") " Nov 24 17:30:20 crc kubenswrapper[4808]: I1124 17:30:20.523501 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thjx9\" (UniqueName: \"kubernetes.io/projected/7a0cb760-04cf-4658-b4ff-1da560789800-kube-api-access-thjx9\") pod \"7a0cb760-04cf-4658-b4ff-1da560789800\" (UID: \"7a0cb760-04cf-4658-b4ff-1da560789800\") " Nov 24 17:30:20 crc kubenswrapper[4808]: I1124 17:30:20.523565 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a0cb760-04cf-4658-b4ff-1da560789800-catalog-content\") pod \"7a0cb760-04cf-4658-b4ff-1da560789800\" (UID: \"7a0cb760-04cf-4658-b4ff-1da560789800\") " Nov 24 17:30:20 crc kubenswrapper[4808]: I1124 17:30:20.523922 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a0cb760-04cf-4658-b4ff-1da560789800-utilities" (OuterVolumeSpecName: "utilities") pod "7a0cb760-04cf-4658-b4ff-1da560789800" (UID: "7a0cb760-04cf-4658-b4ff-1da560789800"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:30:20 crc kubenswrapper[4808]: I1124 17:30:20.524253 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a0cb760-04cf-4658-b4ff-1da560789800-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:20 crc kubenswrapper[4808]: I1124 17:30:20.535270 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a0cb760-04cf-4658-b4ff-1da560789800-kube-api-access-thjx9" (OuterVolumeSpecName: "kube-api-access-thjx9") pod "7a0cb760-04cf-4658-b4ff-1da560789800" (UID: "7a0cb760-04cf-4658-b4ff-1da560789800"). InnerVolumeSpecName "kube-api-access-thjx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:30:20 crc kubenswrapper[4808]: I1124 17:30:20.590734 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a0cb760-04cf-4658-b4ff-1da560789800-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a0cb760-04cf-4658-b4ff-1da560789800" (UID: "7a0cb760-04cf-4658-b4ff-1da560789800"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:30:20 crc kubenswrapper[4808]: I1124 17:30:20.625219 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thjx9\" (UniqueName: \"kubernetes.io/projected/7a0cb760-04cf-4658-b4ff-1da560789800-kube-api-access-thjx9\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:20 crc kubenswrapper[4808]: I1124 17:30:20.625274 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a0cb760-04cf-4658-b4ff-1da560789800-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:20 crc kubenswrapper[4808]: I1124 17:30:20.735810 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5lxz"] Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.065621 4808 generic.go:334] "Generic (PLEG): container finished" podID="7a0cb760-04cf-4658-b4ff-1da560789800" containerID="fe3751e91bad9cee48532b170eb25d59873a1bb344d42adf0031d6cfb4ae765c" exitCode=0 Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.065734 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4pwnf" event={"ID":"7a0cb760-04cf-4658-b4ff-1da560789800","Type":"ContainerDied","Data":"fe3751e91bad9cee48532b170eb25d59873a1bb344d42adf0031d6cfb4ae765c"} Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.065742 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4pwnf" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.065811 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4pwnf" event={"ID":"7a0cb760-04cf-4658-b4ff-1da560789800","Type":"ContainerDied","Data":"7d6d77f6a8ef7dd3c30c52674215fe24c60d8e5c42bd32e8a678659b2bcdc6ea"} Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.065835 4808 scope.go:117] "RemoveContainer" containerID="fe3751e91bad9cee48532b170eb25d59873a1bb344d42adf0031d6cfb4ae765c" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.066580 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v5lxz" podUID="04baaaad-0212-49d6-86af-d498e40e8600" containerName="registry-server" containerID="cri-o://722e267ea35fb19265ac1ce302ba43e6d57d2d65cdcbb5b8c2788786c5410989" gracePeriod=2 Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.136659 4808 scope.go:117] "RemoveContainer" containerID="f7b59228cee8bd2dd1f5862664f50315ac05d1be0d607f8917e40ae1000b115b" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.140549 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4pwnf"] Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.143891 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4pwnf"] Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.196322 4808 scope.go:117] "RemoveContainer" containerID="2094c8fbdfa10f3e12bba5a8aaec3e8bee1fd061bdfc4ab9d85dd7f49f972743" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.220440 4808 scope.go:117] "RemoveContainer" containerID="fe3751e91bad9cee48532b170eb25d59873a1bb344d42adf0031d6cfb4ae765c" Nov 24 17:30:21 crc kubenswrapper[4808]: E1124 17:30:21.222029 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe3751e91bad9cee48532b170eb25d59873a1bb344d42adf0031d6cfb4ae765c\": container with ID starting with fe3751e91bad9cee48532b170eb25d59873a1bb344d42adf0031d6cfb4ae765c not found: ID does not exist" containerID="fe3751e91bad9cee48532b170eb25d59873a1bb344d42adf0031d6cfb4ae765c" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.222107 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe3751e91bad9cee48532b170eb25d59873a1bb344d42adf0031d6cfb4ae765c"} err="failed to get container status \"fe3751e91bad9cee48532b170eb25d59873a1bb344d42adf0031d6cfb4ae765c\": rpc error: code = NotFound desc = could not find container \"fe3751e91bad9cee48532b170eb25d59873a1bb344d42adf0031d6cfb4ae765c\": container with ID starting with fe3751e91bad9cee48532b170eb25d59873a1bb344d42adf0031d6cfb4ae765c not found: ID does not exist" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.222163 4808 scope.go:117] "RemoveContainer" containerID="f7b59228cee8bd2dd1f5862664f50315ac05d1be0d607f8917e40ae1000b115b" Nov 24 17:30:21 crc kubenswrapper[4808]: E1124 17:30:21.222669 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7b59228cee8bd2dd1f5862664f50315ac05d1be0d607f8917e40ae1000b115b\": container with ID starting with f7b59228cee8bd2dd1f5862664f50315ac05d1be0d607f8917e40ae1000b115b not found: ID does not exist" containerID="f7b59228cee8bd2dd1f5862664f50315ac05d1be0d607f8917e40ae1000b115b" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.222733 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7b59228cee8bd2dd1f5862664f50315ac05d1be0d607f8917e40ae1000b115b"} err="failed to get container status \"f7b59228cee8bd2dd1f5862664f50315ac05d1be0d607f8917e40ae1000b115b\": rpc error: code = NotFound desc = could not find container \"f7b59228cee8bd2dd1f5862664f50315ac05d1be0d607f8917e40ae1000b115b\": container with ID starting with f7b59228cee8bd2dd1f5862664f50315ac05d1be0d607f8917e40ae1000b115b not found: ID does not exist" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.222777 4808 scope.go:117] "RemoveContainer" containerID="2094c8fbdfa10f3e12bba5a8aaec3e8bee1fd061bdfc4ab9d85dd7f49f972743" Nov 24 17:30:21 crc kubenswrapper[4808]: E1124 17:30:21.224509 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2094c8fbdfa10f3e12bba5a8aaec3e8bee1fd061bdfc4ab9d85dd7f49f972743\": container with ID starting with 2094c8fbdfa10f3e12bba5a8aaec3e8bee1fd061bdfc4ab9d85dd7f49f972743 not found: ID does not exist" containerID="2094c8fbdfa10f3e12bba5a8aaec3e8bee1fd061bdfc4ab9d85dd7f49f972743" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.224548 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2094c8fbdfa10f3e12bba5a8aaec3e8bee1fd061bdfc4ab9d85dd7f49f972743"} err="failed to get container status \"2094c8fbdfa10f3e12bba5a8aaec3e8bee1fd061bdfc4ab9d85dd7f49f972743\": rpc error: code = NotFound desc = could not find container \"2094c8fbdfa10f3e12bba5a8aaec3e8bee1fd061bdfc4ab9d85dd7f49f972743\": container with ID starting with 2094c8fbdfa10f3e12bba5a8aaec3e8bee1fd061bdfc4ab9d85dd7f49f972743 not found: ID does not exist" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.418180 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.543597 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04baaaad-0212-49d6-86af-d498e40e8600-utilities\") pod \"04baaaad-0212-49d6-86af-d498e40e8600\" (UID: \"04baaaad-0212-49d6-86af-d498e40e8600\") " Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.543709 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m966\" (UniqueName: \"kubernetes.io/projected/04baaaad-0212-49d6-86af-d498e40e8600-kube-api-access-2m966\") pod \"04baaaad-0212-49d6-86af-d498e40e8600\" (UID: \"04baaaad-0212-49d6-86af-d498e40e8600\") " Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.543932 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04baaaad-0212-49d6-86af-d498e40e8600-catalog-content\") pod \"04baaaad-0212-49d6-86af-d498e40e8600\" (UID: \"04baaaad-0212-49d6-86af-d498e40e8600\") " Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.545412 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04baaaad-0212-49d6-86af-d498e40e8600-utilities" (OuterVolumeSpecName: "utilities") pod "04baaaad-0212-49d6-86af-d498e40e8600" (UID: "04baaaad-0212-49d6-86af-d498e40e8600"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.549678 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04baaaad-0212-49d6-86af-d498e40e8600-kube-api-access-2m966" (OuterVolumeSpecName: "kube-api-access-2m966") pod "04baaaad-0212-49d6-86af-d498e40e8600" (UID: "04baaaad-0212-49d6-86af-d498e40e8600"). InnerVolumeSpecName "kube-api-access-2m966". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.554029 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04baaaad-0212-49d6-86af-d498e40e8600-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.554075 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m966\" (UniqueName: \"kubernetes.io/projected/04baaaad-0212-49d6-86af-d498e40e8600-kube-api-access-2m966\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.587432 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04baaaad-0212-49d6-86af-d498e40e8600-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "04baaaad-0212-49d6-86af-d498e40e8600" (UID: "04baaaad-0212-49d6-86af-d498e40e8600"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:30:21 crc kubenswrapper[4808]: I1124 17:30:21.655757 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04baaaad-0212-49d6-86af-d498e40e8600-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.074696 4808 generic.go:334] "Generic (PLEG): container finished" podID="04baaaad-0212-49d6-86af-d498e40e8600" containerID="722e267ea35fb19265ac1ce302ba43e6d57d2d65cdcbb5b8c2788786c5410989" exitCode=0 Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.074774 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5lxz" Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.074836 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5lxz" event={"ID":"04baaaad-0212-49d6-86af-d498e40e8600","Type":"ContainerDied","Data":"722e267ea35fb19265ac1ce302ba43e6d57d2d65cdcbb5b8c2788786c5410989"} Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.074936 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5lxz" event={"ID":"04baaaad-0212-49d6-86af-d498e40e8600","Type":"ContainerDied","Data":"03861f27f5b5679fb91dee5069bde6496967ae4ea12c0c1c3b728e50aec95cd7"} Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.074975 4808 scope.go:117] "RemoveContainer" containerID="722e267ea35fb19265ac1ce302ba43e6d57d2d65cdcbb5b8c2788786c5410989" Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.092949 4808 scope.go:117] "RemoveContainer" containerID="ff96358102311368af9e4d45a5f370a6c053cd6bc57e28ec9b3b756beb134df6" Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.105012 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5lxz"] Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.107966 4808 scope.go:117] "RemoveContainer" containerID="b0fdce55b42a3c7b0f66074ca3d7b206e30f9be44f4aaf79709780802149d99d" Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.109742 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5lxz"] Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.125605 4808 scope.go:117] "RemoveContainer" containerID="722e267ea35fb19265ac1ce302ba43e6d57d2d65cdcbb5b8c2788786c5410989" Nov 24 17:30:22 crc kubenswrapper[4808]: E1124 17:30:22.126238 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"722e267ea35fb19265ac1ce302ba43e6d57d2d65cdcbb5b8c2788786c5410989\": container with ID starting with 722e267ea35fb19265ac1ce302ba43e6d57d2d65cdcbb5b8c2788786c5410989 not found: ID does not exist" containerID="722e267ea35fb19265ac1ce302ba43e6d57d2d65cdcbb5b8c2788786c5410989" Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.126296 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"722e267ea35fb19265ac1ce302ba43e6d57d2d65cdcbb5b8c2788786c5410989"} err="failed to get container status \"722e267ea35fb19265ac1ce302ba43e6d57d2d65cdcbb5b8c2788786c5410989\": rpc error: code = NotFound desc = could not find container \"722e267ea35fb19265ac1ce302ba43e6d57d2d65cdcbb5b8c2788786c5410989\": container with ID starting with 722e267ea35fb19265ac1ce302ba43e6d57d2d65cdcbb5b8c2788786c5410989 not found: ID does not exist" Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.126325 4808 scope.go:117] "RemoveContainer" containerID="ff96358102311368af9e4d45a5f370a6c053cd6bc57e28ec9b3b756beb134df6" Nov 24 17:30:22 crc kubenswrapper[4808]: E1124 17:30:22.126776 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff96358102311368af9e4d45a5f370a6c053cd6bc57e28ec9b3b756beb134df6\": container with ID starting with ff96358102311368af9e4d45a5f370a6c053cd6bc57e28ec9b3b756beb134df6 not found: ID does not exist" containerID="ff96358102311368af9e4d45a5f370a6c053cd6bc57e28ec9b3b756beb134df6" Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.126851 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff96358102311368af9e4d45a5f370a6c053cd6bc57e28ec9b3b756beb134df6"} err="failed to get container status \"ff96358102311368af9e4d45a5f370a6c053cd6bc57e28ec9b3b756beb134df6\": rpc error: code = NotFound desc = could not find container \"ff96358102311368af9e4d45a5f370a6c053cd6bc57e28ec9b3b756beb134df6\": container with ID starting with ff96358102311368af9e4d45a5f370a6c053cd6bc57e28ec9b3b756beb134df6 not found: ID does not exist" Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.126899 4808 scope.go:117] "RemoveContainer" containerID="b0fdce55b42a3c7b0f66074ca3d7b206e30f9be44f4aaf79709780802149d99d" Nov 24 17:30:22 crc kubenswrapper[4808]: E1124 17:30:22.127305 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0fdce55b42a3c7b0f66074ca3d7b206e30f9be44f4aaf79709780802149d99d\": container with ID starting with b0fdce55b42a3c7b0f66074ca3d7b206e30f9be44f4aaf79709780802149d99d not found: ID does not exist" containerID="b0fdce55b42a3c7b0f66074ca3d7b206e30f9be44f4aaf79709780802149d99d" Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.127332 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0fdce55b42a3c7b0f66074ca3d7b206e30f9be44f4aaf79709780802149d99d"} err="failed to get container status \"b0fdce55b42a3c7b0f66074ca3d7b206e30f9be44f4aaf79709780802149d99d\": rpc error: code = NotFound desc = could not find container \"b0fdce55b42a3c7b0f66074ca3d7b206e30f9be44f4aaf79709780802149d99d\": container with ID starting with b0fdce55b42a3c7b0f66074ca3d7b206e30f9be44f4aaf79709780802149d99d not found: ID does not exist" Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.354099 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04baaaad-0212-49d6-86af-d498e40e8600" path="/var/lib/kubelet/pods/04baaaad-0212-49d6-86af-d498e40e8600/volumes" Nov 24 17:30:22 crc kubenswrapper[4808]: I1124 17:30:22.354721 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a0cb760-04cf-4658-b4ff-1da560789800" path="/var/lib/kubelet/pods/7a0cb760-04cf-4658-b4ff-1da560789800/volumes" Nov 24 17:30:23 crc kubenswrapper[4808]: I1124 17:30:23.734352 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bmqk8"] Nov 24 17:30:23 crc kubenswrapper[4808]: I1124 17:30:23.734577 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bmqk8" podUID="234ae5d2-c9a2-470c-9137-f52babb8f9dc" containerName="registry-server" containerID="cri-o://49d65bd2faf950e4b9d8eabe9ec24da5c75c0f0233c1dbbd39d08e003afdb061" gracePeriod=2 Nov 24 17:30:24 crc kubenswrapper[4808]: I1124 17:30:24.948333 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.101623 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/234ae5d2-c9a2-470c-9137-f52babb8f9dc-catalog-content\") pod \"234ae5d2-c9a2-470c-9137-f52babb8f9dc\" (UID: \"234ae5d2-c9a2-470c-9137-f52babb8f9dc\") " Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.101734 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlf7n\" (UniqueName: \"kubernetes.io/projected/234ae5d2-c9a2-470c-9137-f52babb8f9dc-kube-api-access-zlf7n\") pod \"234ae5d2-c9a2-470c-9137-f52babb8f9dc\" (UID: \"234ae5d2-c9a2-470c-9137-f52babb8f9dc\") " Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.101793 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/234ae5d2-c9a2-470c-9137-f52babb8f9dc-utilities\") pod \"234ae5d2-c9a2-470c-9137-f52babb8f9dc\" (UID: \"234ae5d2-c9a2-470c-9137-f52babb8f9dc\") " Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.102348 4808 generic.go:334] "Generic (PLEG): container finished" podID="234ae5d2-c9a2-470c-9137-f52babb8f9dc" containerID="49d65bd2faf950e4b9d8eabe9ec24da5c75c0f0233c1dbbd39d08e003afdb061" exitCode=0 Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.102421 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmqk8" event={"ID":"234ae5d2-c9a2-470c-9137-f52babb8f9dc","Type":"ContainerDied","Data":"49d65bd2faf950e4b9d8eabe9ec24da5c75c0f0233c1dbbd39d08e003afdb061"} Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.102472 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmqk8" event={"ID":"234ae5d2-c9a2-470c-9137-f52babb8f9dc","Type":"ContainerDied","Data":"dbd60cd98e2e1f08507ccbc6e249c0dacde9d58842ac44d5952e272e79fa86c4"} Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.102481 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bmqk8" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.102498 4808 scope.go:117] "RemoveContainer" containerID="49d65bd2faf950e4b9d8eabe9ec24da5c75c0f0233c1dbbd39d08e003afdb061" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.102858 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/234ae5d2-c9a2-470c-9137-f52babb8f9dc-utilities" (OuterVolumeSpecName: "utilities") pod "234ae5d2-c9a2-470c-9137-f52babb8f9dc" (UID: "234ae5d2-c9a2-470c-9137-f52babb8f9dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.109469 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/234ae5d2-c9a2-470c-9137-f52babb8f9dc-kube-api-access-zlf7n" (OuterVolumeSpecName: "kube-api-access-zlf7n") pod "234ae5d2-c9a2-470c-9137-f52babb8f9dc" (UID: "234ae5d2-c9a2-470c-9137-f52babb8f9dc"). InnerVolumeSpecName "kube-api-access-zlf7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.139543 4808 scope.go:117] "RemoveContainer" containerID="08d18bc0d536a68b8f5a0e38ddd17fe6d4a2f7bc8f778c57aa777ceda7cbe6a5" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.156467 4808 scope.go:117] "RemoveContainer" containerID="204caa694017256631117723bd86fde7137f0bb10dbad6efa23ecd6e6893644b" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.173841 4808 scope.go:117] "RemoveContainer" containerID="49d65bd2faf950e4b9d8eabe9ec24da5c75c0f0233c1dbbd39d08e003afdb061" Nov 24 17:30:25 crc kubenswrapper[4808]: E1124 17:30:25.174529 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49d65bd2faf950e4b9d8eabe9ec24da5c75c0f0233c1dbbd39d08e003afdb061\": container with ID starting with 49d65bd2faf950e4b9d8eabe9ec24da5c75c0f0233c1dbbd39d08e003afdb061 not found: ID does not exist" containerID="49d65bd2faf950e4b9d8eabe9ec24da5c75c0f0233c1dbbd39d08e003afdb061" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.174588 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49d65bd2faf950e4b9d8eabe9ec24da5c75c0f0233c1dbbd39d08e003afdb061"} err="failed to get container status \"49d65bd2faf950e4b9d8eabe9ec24da5c75c0f0233c1dbbd39d08e003afdb061\": rpc error: code = NotFound desc = could not find container \"49d65bd2faf950e4b9d8eabe9ec24da5c75c0f0233c1dbbd39d08e003afdb061\": container with ID starting with 49d65bd2faf950e4b9d8eabe9ec24da5c75c0f0233c1dbbd39d08e003afdb061 not found: ID does not exist" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.174629 4808 scope.go:117] "RemoveContainer" containerID="08d18bc0d536a68b8f5a0e38ddd17fe6d4a2f7bc8f778c57aa777ceda7cbe6a5" Nov 24 17:30:25 crc kubenswrapper[4808]: E1124 17:30:25.175139 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08d18bc0d536a68b8f5a0e38ddd17fe6d4a2f7bc8f778c57aa777ceda7cbe6a5\": container with ID starting with 08d18bc0d536a68b8f5a0e38ddd17fe6d4a2f7bc8f778c57aa777ceda7cbe6a5 not found: ID does not exist" containerID="08d18bc0d536a68b8f5a0e38ddd17fe6d4a2f7bc8f778c57aa777ceda7cbe6a5" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.175197 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08d18bc0d536a68b8f5a0e38ddd17fe6d4a2f7bc8f778c57aa777ceda7cbe6a5"} err="failed to get container status \"08d18bc0d536a68b8f5a0e38ddd17fe6d4a2f7bc8f778c57aa777ceda7cbe6a5\": rpc error: code = NotFound desc = could not find container \"08d18bc0d536a68b8f5a0e38ddd17fe6d4a2f7bc8f778c57aa777ceda7cbe6a5\": container with ID starting with 08d18bc0d536a68b8f5a0e38ddd17fe6d4a2f7bc8f778c57aa777ceda7cbe6a5 not found: ID does not exist" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.175226 4808 scope.go:117] "RemoveContainer" containerID="204caa694017256631117723bd86fde7137f0bb10dbad6efa23ecd6e6893644b" Nov 24 17:30:25 crc kubenswrapper[4808]: E1124 17:30:25.175625 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"204caa694017256631117723bd86fde7137f0bb10dbad6efa23ecd6e6893644b\": container with ID starting with 204caa694017256631117723bd86fde7137f0bb10dbad6efa23ecd6e6893644b not found: ID does not exist" containerID="204caa694017256631117723bd86fde7137f0bb10dbad6efa23ecd6e6893644b" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.175669 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"204caa694017256631117723bd86fde7137f0bb10dbad6efa23ecd6e6893644b"} err="failed to get container status \"204caa694017256631117723bd86fde7137f0bb10dbad6efa23ecd6e6893644b\": rpc error: code = NotFound desc = could not find container \"204caa694017256631117723bd86fde7137f0bb10dbad6efa23ecd6e6893644b\": container with ID starting with 204caa694017256631117723bd86fde7137f0bb10dbad6efa23ecd6e6893644b not found: ID does not exist" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.203824 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlf7n\" (UniqueName: \"kubernetes.io/projected/234ae5d2-c9a2-470c-9137-f52babb8f9dc-kube-api-access-zlf7n\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.203869 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/234ae5d2-c9a2-470c-9137-f52babb8f9dc-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.206253 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/234ae5d2-c9a2-470c-9137-f52babb8f9dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "234ae5d2-c9a2-470c-9137-f52babb8f9dc" (UID: "234ae5d2-c9a2-470c-9137-f52babb8f9dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.305203 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/234ae5d2-c9a2-470c-9137-f52babb8f9dc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.445942 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bmqk8"] Nov 24 17:30:25 crc kubenswrapper[4808]: I1124 17:30:25.448575 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bmqk8"] Nov 24 17:30:26 crc kubenswrapper[4808]: I1124 17:30:26.253992 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:30:26 crc kubenswrapper[4808]: I1124 17:30:26.254147 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:30:26 crc kubenswrapper[4808]: I1124 17:30:26.315243 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:30:26 crc kubenswrapper[4808]: I1124 17:30:26.355685 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="234ae5d2-c9a2-470c-9137-f52babb8f9dc" path="/var/lib/kubelet/pods/234ae5d2-c9a2-470c-9137-f52babb8f9dc/volumes" Nov 24 17:30:27 crc kubenswrapper[4808]: I1124 17:30:27.155859 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:30:28 crc kubenswrapper[4808]: I1124 17:30:28.281586 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:30:29 crc kubenswrapper[4808]: I1124 17:30:29.153053 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-b5tvs"] Nov 24 17:30:29 crc kubenswrapper[4808]: I1124 17:30:29.497350 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:30:29 crc kubenswrapper[4808]: I1124 17:30:29.543130 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:30:36 crc kubenswrapper[4808]: I1124 17:30:36.523440 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:30:36 crc kubenswrapper[4808]: I1124 17:30:36.524523 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:30:36 crc kubenswrapper[4808]: I1124 17:30:36.524598 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:30:37 crc kubenswrapper[4808]: I1124 17:30:37.179406 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8"} pod="openshift-machine-config-operator/machine-config-daemon-ht87w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:30:37 crc kubenswrapper[4808]: I1124 17:30:37.179490 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" containerID="cri-o://bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8" gracePeriod=600 Nov 24 17:30:38 crc kubenswrapper[4808]: I1124 17:30:38.188410 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerID="bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8" exitCode=0 Nov 24 17:30:38 crc kubenswrapper[4808]: I1124 17:30:38.188507 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerDied","Data":"bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8"} Nov 24 17:30:38 crc kubenswrapper[4808]: I1124 17:30:38.189038 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"4b9523f5392391855bf06e84ed62910c36d9c20fb3584ca98f4425fec2b9f41c"} Nov 24 17:30:54 crc kubenswrapper[4808]: I1124 17:30:54.194688 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" podUID="0416a1dd-4c58-4d18-b6f3-404e4c642e6b" containerName="oauth-openshift" containerID="cri-o://ba947ce528c834968431673cff84469bc626c08c90654a64e43ad10b0953a273" gracePeriod=15 Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.094780 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135347 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5db8794bf8-8xm7d"] Nov 24 17:30:55 crc kubenswrapper[4808]: E1124 17:30:55.135614 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234ae5d2-c9a2-470c-9137-f52babb8f9dc" containerName="extract-utilities" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135628 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="234ae5d2-c9a2-470c-9137-f52babb8f9dc" containerName="extract-utilities" Nov 24 17:30:55 crc kubenswrapper[4808]: E1124 17:30:55.135640 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234ae5d2-c9a2-470c-9137-f52babb8f9dc" containerName="registry-server" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135646 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="234ae5d2-c9a2-470c-9137-f52babb8f9dc" containerName="registry-server" Nov 24 17:30:55 crc kubenswrapper[4808]: E1124 17:30:55.135661 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a0cb760-04cf-4658-b4ff-1da560789800" containerName="extract-utilities" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135667 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a0cb760-04cf-4658-b4ff-1da560789800" containerName="extract-utilities" Nov 24 17:30:55 crc kubenswrapper[4808]: E1124 17:30:55.135677 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04baaaad-0212-49d6-86af-d498e40e8600" containerName="registry-server" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135683 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="04baaaad-0212-49d6-86af-d498e40e8600" containerName="registry-server" Nov 24 17:30:55 crc kubenswrapper[4808]: E1124 17:30:55.135691 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a0cb760-04cf-4658-b4ff-1da560789800" containerName="extract-content" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135697 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a0cb760-04cf-4658-b4ff-1da560789800" containerName="extract-content" Nov 24 17:30:55 crc kubenswrapper[4808]: E1124 17:30:55.135705 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234ae5d2-c9a2-470c-9137-f52babb8f9dc" containerName="extract-content" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135711 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="234ae5d2-c9a2-470c-9137-f52babb8f9dc" containerName="extract-content" Nov 24 17:30:55 crc kubenswrapper[4808]: E1124 17:30:55.135719 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a0cb760-04cf-4658-b4ff-1da560789800" containerName="registry-server" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135725 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a0cb760-04cf-4658-b4ff-1da560789800" containerName="registry-server" Nov 24 17:30:55 crc kubenswrapper[4808]: E1124 17:30:55.135733 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0416a1dd-4c58-4d18-b6f3-404e4c642e6b" containerName="oauth-openshift" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135739 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="0416a1dd-4c58-4d18-b6f3-404e4c642e6b" containerName="oauth-openshift" Nov 24 17:30:55 crc kubenswrapper[4808]: E1124 17:30:55.135747 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04baaaad-0212-49d6-86af-d498e40e8600" containerName="extract-content" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135754 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="04baaaad-0212-49d6-86af-d498e40e8600" containerName="extract-content" Nov 24 17:30:55 crc kubenswrapper[4808]: E1124 17:30:55.135763 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04baaaad-0212-49d6-86af-d498e40e8600" containerName="extract-utilities" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135771 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="04baaaad-0212-49d6-86af-d498e40e8600" containerName="extract-utilities" Nov 24 17:30:55 crc kubenswrapper[4808]: E1124 17:30:55.135788 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" containerName="extract-utilities" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135800 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" containerName="extract-utilities" Nov 24 17:30:55 crc kubenswrapper[4808]: E1124 17:30:55.135811 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" containerName="registry-server" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135820 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" containerName="registry-server" Nov 24 17:30:55 crc kubenswrapper[4808]: E1124 17:30:55.135831 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" containerName="extract-content" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135839 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" containerName="extract-content" Nov 24 17:30:55 crc kubenswrapper[4808]: E1124 17:30:55.135851 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fee761ab-3ef8-4ee5-9bea-59395abe3fbc" containerName="collect-profiles" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135879 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="fee761ab-3ef8-4ee5-9bea-59395abe3fbc" containerName="collect-profiles" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135978 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="234ae5d2-c9a2-470c-9137-f52babb8f9dc" containerName="registry-server" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135987 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="fee761ab-3ef8-4ee5-9bea-59395abe3fbc" containerName="collect-profiles" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.135997 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="04baaaad-0212-49d6-86af-d498e40e8600" containerName="registry-server" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.136003 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a0cb760-04cf-4658-b4ff-1da560789800" containerName="registry-server" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.136030 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="0416a1dd-4c58-4d18-b6f3-404e4c642e6b" containerName="oauth-openshift" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.136036 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1ef2dab-8530-43cf-a72e-22cb8ac8e1a3" containerName="registry-server" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.136510 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.149991 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5db8794bf8-8xm7d"] Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.229128 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-ocp-branding-template\") pod \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.229290 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftdwt\" (UniqueName: \"kubernetes.io/projected/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-kube-api-access-ftdwt\") pod \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.229322 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-cliconfig\") pod \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.229350 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-router-certs\") pod \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.229371 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-trusted-ca-bundle\") pod \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.229426 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-serving-cert\") pod \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.229458 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-provider-selection\") pod \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.229487 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-error\") pod \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.229509 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-login\") pod \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.229535 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-idp-0-file-data\") pod \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.229567 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-service-ca\") pod \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.229585 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-audit-policies\") pod \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.229611 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-audit-dir\") pod \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.229631 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-session\") pod \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\" (UID: \"0416a1dd-4c58-4d18-b6f3-404e4c642e6b\") " Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.229993 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.230060 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-session\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.230083 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9nvg\" (UniqueName: \"kubernetes.io/projected/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-kube-api-access-d9nvg\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.230103 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.230130 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.230155 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-audit-dir\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.230173 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-router-certs\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.230280 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-service-ca\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.230582 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.232158 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "0416a1dd-4c58-4d18-b6f3-404e4c642e6b" (UID: "0416a1dd-4c58-4d18-b6f3-404e4c642e6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.232197 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-user-template-error\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.232257 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "0416a1dd-4c58-4d18-b6f3-404e4c642e6b" (UID: "0416a1dd-4c58-4d18-b6f3-404e4c642e6b"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.232299 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "0416a1dd-4c58-4d18-b6f3-404e4c642e6b" (UID: "0416a1dd-4c58-4d18-b6f3-404e4c642e6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.232414 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-user-template-login\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.232461 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.232477 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "0416a1dd-4c58-4d18-b6f3-404e4c642e6b" (UID: "0416a1dd-4c58-4d18-b6f3-404e4c642e6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.232696 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.232733 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-audit-policies\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.232911 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.232927 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "0416a1dd-4c58-4d18-b6f3-404e4c642e6b" (UID: "0416a1dd-4c58-4d18-b6f3-404e4c642e6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.232937 4808 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.232987 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.233212 4808 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.238412 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "0416a1dd-4c58-4d18-b6f3-404e4c642e6b" (UID: "0416a1dd-4c58-4d18-b6f3-404e4c642e6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.239118 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "0416a1dd-4c58-4d18-b6f3-404e4c642e6b" (UID: "0416a1dd-4c58-4d18-b6f3-404e4c642e6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.239236 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "0416a1dd-4c58-4d18-b6f3-404e4c642e6b" (UID: "0416a1dd-4c58-4d18-b6f3-404e4c642e6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.239560 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "0416a1dd-4c58-4d18-b6f3-404e4c642e6b" (UID: "0416a1dd-4c58-4d18-b6f3-404e4c642e6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.239823 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-kube-api-access-ftdwt" (OuterVolumeSpecName: "kube-api-access-ftdwt") pod "0416a1dd-4c58-4d18-b6f3-404e4c642e6b" (UID: "0416a1dd-4c58-4d18-b6f3-404e4c642e6b"). InnerVolumeSpecName "kube-api-access-ftdwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.239909 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "0416a1dd-4c58-4d18-b6f3-404e4c642e6b" (UID: "0416a1dd-4c58-4d18-b6f3-404e4c642e6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.242435 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "0416a1dd-4c58-4d18-b6f3-404e4c642e6b" (UID: "0416a1dd-4c58-4d18-b6f3-404e4c642e6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.242622 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "0416a1dd-4c58-4d18-b6f3-404e4c642e6b" (UID: "0416a1dd-4c58-4d18-b6f3-404e4c642e6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.242743 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "0416a1dd-4c58-4d18-b6f3-404e4c642e6b" (UID: "0416a1dd-4c58-4d18-b6f3-404e4c642e6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.311843 4808 generic.go:334] "Generic (PLEG): container finished" podID="0416a1dd-4c58-4d18-b6f3-404e4c642e6b" containerID="ba947ce528c834968431673cff84469bc626c08c90654a64e43ad10b0953a273" exitCode=0 Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.311905 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" event={"ID":"0416a1dd-4c58-4d18-b6f3-404e4c642e6b","Type":"ContainerDied","Data":"ba947ce528c834968431673cff84469bc626c08c90654a64e43ad10b0953a273"} Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.311973 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" event={"ID":"0416a1dd-4c58-4d18-b6f3-404e4c642e6b","Type":"ContainerDied","Data":"1a81f64f0b332f7acf2ed04c8759611e350f37d94a302dfd8742848c16388436"} Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.312002 4808 scope.go:117] "RemoveContainer" containerID="ba947ce528c834968431673cff84469bc626c08c90654a64e43ad10b0953a273" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.313060 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-b5tvs" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.334709 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-service-ca\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.334795 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.334832 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-user-template-error\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.334872 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-user-template-login\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.334923 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335061 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335098 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-audit-policies\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335127 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335160 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-session\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335185 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9nvg\" (UniqueName: \"kubernetes.io/projected/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-kube-api-access-d9nvg\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335212 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335251 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335291 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-audit-dir\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335321 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-router-certs\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335375 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335613 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335630 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335645 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335655 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335669 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335681 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335692 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftdwt\" (UniqueName: \"kubernetes.io/projected/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-kube-api-access-ftdwt\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335703 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335713 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0416a1dd-4c58-4d18-b6f3-404e4c642e6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.335770 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.336238 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-service-ca\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.336351 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-audit-dir\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.336471 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-audit-policies\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.337283 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.339350 4808 scope.go:117] "RemoveContainer" containerID="ba947ce528c834968431673cff84469bc626c08c90654a64e43ad10b0953a273" Nov 24 17:30:55 crc kubenswrapper[4808]: E1124 17:30:55.340598 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba947ce528c834968431673cff84469bc626c08c90654a64e43ad10b0953a273\": container with ID starting with ba947ce528c834968431673cff84469bc626c08c90654a64e43ad10b0953a273 not found: ID does not exist" containerID="ba947ce528c834968431673cff84469bc626c08c90654a64e43ad10b0953a273" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.340641 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba947ce528c834968431673cff84469bc626c08c90654a64e43ad10b0953a273"} err="failed to get container status \"ba947ce528c834968431673cff84469bc626c08c90654a64e43ad10b0953a273\": rpc error: code = NotFound desc = could not find container \"ba947ce528c834968431673cff84469bc626c08c90654a64e43ad10b0953a273\": container with ID starting with ba947ce528c834968431673cff84469bc626c08c90654a64e43ad10b0953a273 not found: ID does not exist" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.344163 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.345768 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-session\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.347604 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.349299 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-user-template-login\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.350255 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-router-certs\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.351341 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-b5tvs"] Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.351816 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.359241 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-b5tvs"] Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.359264 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-user-template-error\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.361270 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9nvg\" (UniqueName: \"kubernetes.io/projected/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-kube-api-access-d9nvg\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.361909 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2c2a2011-3e5b-4713-a8d1-ae3412faecc7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5db8794bf8-8xm7d\" (UID: \"2c2a2011-3e5b-4713-a8d1-ae3412faecc7\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.458623 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:55 crc kubenswrapper[4808]: I1124 17:30:55.650049 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5db8794bf8-8xm7d"] Nov 24 17:30:56 crc kubenswrapper[4808]: I1124 17:30:56.321399 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" event={"ID":"2c2a2011-3e5b-4713-a8d1-ae3412faecc7","Type":"ContainerStarted","Data":"65a71027fb154d2a5cd7981af1949d43d5ac90961927680eccde0af4212821e6"} Nov 24 17:30:56 crc kubenswrapper[4808]: I1124 17:30:56.321789 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" event={"ID":"2c2a2011-3e5b-4713-a8d1-ae3412faecc7","Type":"ContainerStarted","Data":"4489ae2dad3fed6c9c1c9c6aa9ca12b239669f81909f2495fdbdcc3fd3f81e30"} Nov 24 17:30:56 crc kubenswrapper[4808]: I1124 17:30:56.322239 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:56 crc kubenswrapper[4808]: I1124 17:30:56.331032 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" Nov 24 17:30:56 crc kubenswrapper[4808]: I1124 17:30:56.350556 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5db8794bf8-8xm7d" podStartSLOduration=27.350522898 podStartE2EDuration="27.350522898s" podCreationTimestamp="2025-11-24 17:30:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:30:56.347792556 +0000 UTC m=+248.945460358" watchObservedRunningTime="2025-11-24 17:30:56.350522898 +0000 UTC m=+248.948190720" Nov 24 17:30:56 crc kubenswrapper[4808]: I1124 17:30:56.357447 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0416a1dd-4c58-4d18-b6f3-404e4c642e6b" path="/var/lib/kubelet/pods/0416a1dd-4c58-4d18-b6f3-404e4c642e6b/volumes" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.106762 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fwnlt"] Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.108123 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fwnlt" podUID="f11d89a2-be43-47fd-ae09-57fcd7df9820" containerName="registry-server" containerID="cri-o://232dc4691cf01a2e914d3b89c3027a795a276f026a1fd2558510f1e52e155a22" gracePeriod=30 Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.122327 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mlmvz"] Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.122667 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mlmvz" podUID="a3f2246d-a6bb-463f-9a38-2d7144079eb6" containerName="registry-server" containerID="cri-o://c2e447480116b74cbcb702d927fd337acfa8eb284076997d3fe05668eb7853ac" gracePeriod=30 Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.128380 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2v2lb"] Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.128610 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" podUID="94a5dac5-a893-481f-ab4e-bfbd4106b8c6" containerName="marketplace-operator" containerID="cri-o://8b1591ff5b64fd5ed0c2ae4b6e0e871c54201749042134594259363a5cf31f3e" gracePeriod=30 Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.141789 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-96cn9"] Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.142054 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-96cn9" podUID="cde9313c-9316-40dc-8427-a8683ce67b5b" containerName="registry-server" containerID="cri-o://69001147dd08c93f8adef7b397ecfe5c5cdd8992418d86e2a9624b691cdeee11" gracePeriod=30 Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.150070 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5wdrd"] Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.150414 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5wdrd" podUID="b005cb77-3bba-44a5-8446-f6fcc5053336" containerName="registry-server" containerID="cri-o://2d7a3cb59965a53140d64612d3ab007f76ea1382eeef72510e0d469dee1c791a" gracePeriod=30 Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.165257 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-555rd"] Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.166717 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-555rd" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.179665 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-555rd"] Nov 24 17:31:06 crc kubenswrapper[4808]: E1124 17:31:06.254962 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 232dc4691cf01a2e914d3b89c3027a795a276f026a1fd2558510f1e52e155a22 is running failed: container process not found" containerID="232dc4691cf01a2e914d3b89c3027a795a276f026a1fd2558510f1e52e155a22" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 17:31:06 crc kubenswrapper[4808]: E1124 17:31:06.255417 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 232dc4691cf01a2e914d3b89c3027a795a276f026a1fd2558510f1e52e155a22 is running failed: container process not found" containerID="232dc4691cf01a2e914d3b89c3027a795a276f026a1fd2558510f1e52e155a22" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 17:31:06 crc kubenswrapper[4808]: E1124 17:31:06.257469 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 232dc4691cf01a2e914d3b89c3027a795a276f026a1fd2558510f1e52e155a22 is running failed: container process not found" containerID="232dc4691cf01a2e914d3b89c3027a795a276f026a1fd2558510f1e52e155a22" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 17:31:06 crc kubenswrapper[4808]: E1124 17:31:06.257503 4808 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 232dc4691cf01a2e914d3b89c3027a795a276f026a1fd2558510f1e52e155a22 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-fwnlt" podUID="f11d89a2-be43-47fd-ae09-57fcd7df9820" containerName="registry-server" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.298622 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcdth\" (UniqueName: \"kubernetes.io/projected/6ae01b29-6123-4219-b106-b8e55e83cb7f-kube-api-access-wcdth\") pod \"marketplace-operator-79b997595-555rd\" (UID: \"6ae01b29-6123-4219-b106-b8e55e83cb7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-555rd" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.299190 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6ae01b29-6123-4219-b106-b8e55e83cb7f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-555rd\" (UID: \"6ae01b29-6123-4219-b106-b8e55e83cb7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-555rd" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.299237 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6ae01b29-6123-4219-b106-b8e55e83cb7f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-555rd\" (UID: \"6ae01b29-6123-4219-b106-b8e55e83cb7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-555rd" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.380045 4808 generic.go:334] "Generic (PLEG): container finished" podID="cde9313c-9316-40dc-8427-a8683ce67b5b" containerID="69001147dd08c93f8adef7b397ecfe5c5cdd8992418d86e2a9624b691cdeee11" exitCode=0 Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.380128 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96cn9" event={"ID":"cde9313c-9316-40dc-8427-a8683ce67b5b","Type":"ContainerDied","Data":"69001147dd08c93f8adef7b397ecfe5c5cdd8992418d86e2a9624b691cdeee11"} Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.386854 4808 generic.go:334] "Generic (PLEG): container finished" podID="94a5dac5-a893-481f-ab4e-bfbd4106b8c6" containerID="8b1591ff5b64fd5ed0c2ae4b6e0e871c54201749042134594259363a5cf31f3e" exitCode=0 Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.386923 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" event={"ID":"94a5dac5-a893-481f-ab4e-bfbd4106b8c6","Type":"ContainerDied","Data":"8b1591ff5b64fd5ed0c2ae4b6e0e871c54201749042134594259363a5cf31f3e"} Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.389287 4808 generic.go:334] "Generic (PLEG): container finished" podID="f11d89a2-be43-47fd-ae09-57fcd7df9820" containerID="232dc4691cf01a2e914d3b89c3027a795a276f026a1fd2558510f1e52e155a22" exitCode=0 Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.389340 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fwnlt" event={"ID":"f11d89a2-be43-47fd-ae09-57fcd7df9820","Type":"ContainerDied","Data":"232dc4691cf01a2e914d3b89c3027a795a276f026a1fd2558510f1e52e155a22"} Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.393502 4808 generic.go:334] "Generic (PLEG): container finished" podID="a3f2246d-a6bb-463f-9a38-2d7144079eb6" containerID="c2e447480116b74cbcb702d927fd337acfa8eb284076997d3fe05668eb7853ac" exitCode=0 Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.393598 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlmvz" event={"ID":"a3f2246d-a6bb-463f-9a38-2d7144079eb6","Type":"ContainerDied","Data":"c2e447480116b74cbcb702d927fd337acfa8eb284076997d3fe05668eb7853ac"} Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.396759 4808 generic.go:334] "Generic (PLEG): container finished" podID="b005cb77-3bba-44a5-8446-f6fcc5053336" containerID="2d7a3cb59965a53140d64612d3ab007f76ea1382eeef72510e0d469dee1c791a" exitCode=0 Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.396789 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5wdrd" event={"ID":"b005cb77-3bba-44a5-8446-f6fcc5053336","Type":"ContainerDied","Data":"2d7a3cb59965a53140d64612d3ab007f76ea1382eeef72510e0d469dee1c791a"} Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.401215 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcdth\" (UniqueName: \"kubernetes.io/projected/6ae01b29-6123-4219-b106-b8e55e83cb7f-kube-api-access-wcdth\") pod \"marketplace-operator-79b997595-555rd\" (UID: \"6ae01b29-6123-4219-b106-b8e55e83cb7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-555rd" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.401281 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6ae01b29-6123-4219-b106-b8e55e83cb7f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-555rd\" (UID: \"6ae01b29-6123-4219-b106-b8e55e83cb7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-555rd" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.401316 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6ae01b29-6123-4219-b106-b8e55e83cb7f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-555rd\" (UID: \"6ae01b29-6123-4219-b106-b8e55e83cb7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-555rd" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.403157 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6ae01b29-6123-4219-b106-b8e55e83cb7f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-555rd\" (UID: \"6ae01b29-6123-4219-b106-b8e55e83cb7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-555rd" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.412450 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6ae01b29-6123-4219-b106-b8e55e83cb7f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-555rd\" (UID: \"6ae01b29-6123-4219-b106-b8e55e83cb7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-555rd" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.419638 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcdth\" (UniqueName: \"kubernetes.io/projected/6ae01b29-6123-4219-b106-b8e55e83cb7f-kube-api-access-wcdth\") pod \"marketplace-operator-79b997595-555rd\" (UID: \"6ae01b29-6123-4219-b106-b8e55e83cb7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-555rd" Nov 24 17:31:06 crc kubenswrapper[4808]: E1124 17:31:06.438613 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c2e447480116b74cbcb702d927fd337acfa8eb284076997d3fe05668eb7853ac is running failed: container process not found" containerID="c2e447480116b74cbcb702d927fd337acfa8eb284076997d3fe05668eb7853ac" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 17:31:06 crc kubenswrapper[4808]: E1124 17:31:06.439153 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c2e447480116b74cbcb702d927fd337acfa8eb284076997d3fe05668eb7853ac is running failed: container process not found" containerID="c2e447480116b74cbcb702d927fd337acfa8eb284076997d3fe05668eb7853ac" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 17:31:06 crc kubenswrapper[4808]: E1124 17:31:06.439483 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c2e447480116b74cbcb702d927fd337acfa8eb284076997d3fe05668eb7853ac is running failed: container process not found" containerID="c2e447480116b74cbcb702d927fd337acfa8eb284076997d3fe05668eb7853ac" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 17:31:06 crc kubenswrapper[4808]: E1124 17:31:06.439509 4808 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c2e447480116b74cbcb702d927fd337acfa8eb284076997d3fe05668eb7853ac is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-mlmvz" podUID="a3f2246d-a6bb-463f-9a38-2d7144079eb6" containerName="registry-server" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.608778 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-555rd" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.613224 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.634251 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.634841 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.652301 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.660759 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.711326 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f2246d-a6bb-463f-9a38-2d7144079eb6-utilities\") pod \"a3f2246d-a6bb-463f-9a38-2d7144079eb6\" (UID: \"a3f2246d-a6bb-463f-9a38-2d7144079eb6\") " Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.711406 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-marketplace-operator-metrics\") pod \"94a5dac5-a893-481f-ab4e-bfbd4106b8c6\" (UID: \"94a5dac5-a893-481f-ab4e-bfbd4106b8c6\") " Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.711451 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4g78\" (UniqueName: \"kubernetes.io/projected/f11d89a2-be43-47fd-ae09-57fcd7df9820-kube-api-access-w4g78\") pod \"f11d89a2-be43-47fd-ae09-57fcd7df9820\" (UID: \"f11d89a2-be43-47fd-ae09-57fcd7df9820\") " Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.711473 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f11d89a2-be43-47fd-ae09-57fcd7df9820-catalog-content\") pod \"f11d89a2-be43-47fd-ae09-57fcd7df9820\" (UID: \"f11d89a2-be43-47fd-ae09-57fcd7df9820\") " Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.711502 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b005cb77-3bba-44a5-8446-f6fcc5053336-catalog-content\") pod \"b005cb77-3bba-44a5-8446-f6fcc5053336\" (UID: \"b005cb77-3bba-44a5-8446-f6fcc5053336\") " Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.711537 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f11d89a2-be43-47fd-ae09-57fcd7df9820-utilities\") pod \"f11d89a2-be43-47fd-ae09-57fcd7df9820\" (UID: \"f11d89a2-be43-47fd-ae09-57fcd7df9820\") " Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.711563 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nblks\" (UniqueName: \"kubernetes.io/projected/a3f2246d-a6bb-463f-9a38-2d7144079eb6-kube-api-access-nblks\") pod \"a3f2246d-a6bb-463f-9a38-2d7144079eb6\" (UID: \"a3f2246d-a6bb-463f-9a38-2d7144079eb6\") " Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.711586 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b005cb77-3bba-44a5-8446-f6fcc5053336-utilities\") pod \"b005cb77-3bba-44a5-8446-f6fcc5053336\" (UID: \"b005cb77-3bba-44a5-8446-f6fcc5053336\") " Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.711626 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98lb9\" (UniqueName: \"kubernetes.io/projected/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-kube-api-access-98lb9\") pod \"94a5dac5-a893-481f-ab4e-bfbd4106b8c6\" (UID: \"94a5dac5-a893-481f-ab4e-bfbd4106b8c6\") " Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.711654 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cde9313c-9316-40dc-8427-a8683ce67b5b-catalog-content\") pod \"cde9313c-9316-40dc-8427-a8683ce67b5b\" (UID: \"cde9313c-9316-40dc-8427-a8683ce67b5b\") " Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.711687 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f2246d-a6bb-463f-9a38-2d7144079eb6-catalog-content\") pod \"a3f2246d-a6bb-463f-9a38-2d7144079eb6\" (UID: \"a3f2246d-a6bb-463f-9a38-2d7144079eb6\") " Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.711716 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2r6w\" (UniqueName: \"kubernetes.io/projected/cde9313c-9316-40dc-8427-a8683ce67b5b-kube-api-access-v2r6w\") pod \"cde9313c-9316-40dc-8427-a8683ce67b5b\" (UID: \"cde9313c-9316-40dc-8427-a8683ce67b5b\") " Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.711749 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-marketplace-trusted-ca\") pod \"94a5dac5-a893-481f-ab4e-bfbd4106b8c6\" (UID: \"94a5dac5-a893-481f-ab4e-bfbd4106b8c6\") " Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.711782 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cde9313c-9316-40dc-8427-a8683ce67b5b-utilities\") pod \"cde9313c-9316-40dc-8427-a8683ce67b5b\" (UID: \"cde9313c-9316-40dc-8427-a8683ce67b5b\") " Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.711875 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqn4d\" (UniqueName: \"kubernetes.io/projected/b005cb77-3bba-44a5-8446-f6fcc5053336-kube-api-access-qqn4d\") pod \"b005cb77-3bba-44a5-8446-f6fcc5053336\" (UID: \"b005cb77-3bba-44a5-8446-f6fcc5053336\") " Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.718682 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3f2246d-a6bb-463f-9a38-2d7144079eb6-kube-api-access-nblks" (OuterVolumeSpecName: "kube-api-access-nblks") pod "a3f2246d-a6bb-463f-9a38-2d7144079eb6" (UID: "a3f2246d-a6bb-463f-9a38-2d7144079eb6"). InnerVolumeSpecName "kube-api-access-nblks". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.719247 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b005cb77-3bba-44a5-8446-f6fcc5053336-kube-api-access-qqn4d" (OuterVolumeSpecName: "kube-api-access-qqn4d") pod "b005cb77-3bba-44a5-8446-f6fcc5053336" (UID: "b005cb77-3bba-44a5-8446-f6fcc5053336"). InnerVolumeSpecName "kube-api-access-qqn4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.719517 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f11d89a2-be43-47fd-ae09-57fcd7df9820-utilities" (OuterVolumeSpecName: "utilities") pod "f11d89a2-be43-47fd-ae09-57fcd7df9820" (UID: "f11d89a2-be43-47fd-ae09-57fcd7df9820"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.720683 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cde9313c-9316-40dc-8427-a8683ce67b5b-utilities" (OuterVolumeSpecName: "utilities") pod "cde9313c-9316-40dc-8427-a8683ce67b5b" (UID: "cde9313c-9316-40dc-8427-a8683ce67b5b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.721827 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b005cb77-3bba-44a5-8446-f6fcc5053336-utilities" (OuterVolumeSpecName: "utilities") pod "b005cb77-3bba-44a5-8446-f6fcc5053336" (UID: "b005cb77-3bba-44a5-8446-f6fcc5053336"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.722368 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "94a5dac5-a893-481f-ab4e-bfbd4106b8c6" (UID: "94a5dac5-a893-481f-ab4e-bfbd4106b8c6"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.723225 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f11d89a2-be43-47fd-ae09-57fcd7df9820-kube-api-access-w4g78" (OuterVolumeSpecName: "kube-api-access-w4g78") pod "f11d89a2-be43-47fd-ae09-57fcd7df9820" (UID: "f11d89a2-be43-47fd-ae09-57fcd7df9820"). InnerVolumeSpecName "kube-api-access-w4g78". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.723602 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3f2246d-a6bb-463f-9a38-2d7144079eb6-utilities" (OuterVolumeSpecName: "utilities") pod "a3f2246d-a6bb-463f-9a38-2d7144079eb6" (UID: "a3f2246d-a6bb-463f-9a38-2d7144079eb6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.731198 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "94a5dac5-a893-481f-ab4e-bfbd4106b8c6" (UID: "94a5dac5-a893-481f-ab4e-bfbd4106b8c6"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.731221 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cde9313c-9316-40dc-8427-a8683ce67b5b-kube-api-access-v2r6w" (OuterVolumeSpecName: "kube-api-access-v2r6w") pod "cde9313c-9316-40dc-8427-a8683ce67b5b" (UID: "cde9313c-9316-40dc-8427-a8683ce67b5b"). InnerVolumeSpecName "kube-api-access-v2r6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.731255 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-kube-api-access-98lb9" (OuterVolumeSpecName: "kube-api-access-98lb9") pod "94a5dac5-a893-481f-ab4e-bfbd4106b8c6" (UID: "94a5dac5-a893-481f-ab4e-bfbd4106b8c6"). InnerVolumeSpecName "kube-api-access-98lb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.747593 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cde9313c-9316-40dc-8427-a8683ce67b5b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cde9313c-9316-40dc-8427-a8683ce67b5b" (UID: "cde9313c-9316-40dc-8427-a8683ce67b5b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.813242 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cde9313c-9316-40dc-8427-a8683ce67b5b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.813287 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqn4d\" (UniqueName: \"kubernetes.io/projected/b005cb77-3bba-44a5-8446-f6fcc5053336-kube-api-access-qqn4d\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.813302 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f2246d-a6bb-463f-9a38-2d7144079eb6-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.813314 4808 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.813326 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4g78\" (UniqueName: \"kubernetes.io/projected/f11d89a2-be43-47fd-ae09-57fcd7df9820-kube-api-access-w4g78\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.813350 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f11d89a2-be43-47fd-ae09-57fcd7df9820-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.813362 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nblks\" (UniqueName: \"kubernetes.io/projected/a3f2246d-a6bb-463f-9a38-2d7144079eb6-kube-api-access-nblks\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.813371 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b005cb77-3bba-44a5-8446-f6fcc5053336-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.813379 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98lb9\" (UniqueName: \"kubernetes.io/projected/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-kube-api-access-98lb9\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.813390 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cde9313c-9316-40dc-8427-a8683ce67b5b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.813399 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2r6w\" (UniqueName: \"kubernetes.io/projected/cde9313c-9316-40dc-8427-a8683ce67b5b-kube-api-access-v2r6w\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.813407 4808 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94a5dac5-a893-481f-ab4e-bfbd4106b8c6-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.820886 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f11d89a2-be43-47fd-ae09-57fcd7df9820-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f11d89a2-be43-47fd-ae09-57fcd7df9820" (UID: "f11d89a2-be43-47fd-ae09-57fcd7df9820"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.834031 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b005cb77-3bba-44a5-8446-f6fcc5053336-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b005cb77-3bba-44a5-8446-f6fcc5053336" (UID: "b005cb77-3bba-44a5-8446-f6fcc5053336"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.841606 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3f2246d-a6bb-463f-9a38-2d7144079eb6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3f2246d-a6bb-463f-9a38-2d7144079eb6" (UID: "a3f2246d-a6bb-463f-9a38-2d7144079eb6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.874187 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-555rd"] Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.915698 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f11d89a2-be43-47fd-ae09-57fcd7df9820-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.915767 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b005cb77-3bba-44a5-8446-f6fcc5053336-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:06 crc kubenswrapper[4808]: I1124 17:31:06.915795 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f2246d-a6bb-463f-9a38-2d7144079eb6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.405269 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlmvz" event={"ID":"a3f2246d-a6bb-463f-9a38-2d7144079eb6","Type":"ContainerDied","Data":"3f093bdb719ed3e403f63954a18527a1ceffb807e19e09157dce1bea08862407"} Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.405421 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mlmvz" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.405791 4808 scope.go:117] "RemoveContainer" containerID="c2e447480116b74cbcb702d927fd337acfa8eb284076997d3fe05668eb7853ac" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.407608 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5wdrd" event={"ID":"b005cb77-3bba-44a5-8446-f6fcc5053336","Type":"ContainerDied","Data":"dc3a758c84a2c6ffcec0d792ccb1cd3784c3848fa347c5aeaa071060afc79ce8"} Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.407784 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5wdrd" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.408762 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-555rd" event={"ID":"6ae01b29-6123-4219-b106-b8e55e83cb7f","Type":"ContainerStarted","Data":"1f45a467ff98bd6255a3ddfa7beccd8052b3d1a2dba80c6841c8b27875a353ff"} Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.409414 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-555rd" event={"ID":"6ae01b29-6123-4219-b106-b8e55e83cb7f","Type":"ContainerStarted","Data":"1c0d13196936f76a97ab7a593b8e4f3b189eb9bab6c66d8094a59daf12816624"} Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.409470 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-555rd" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.411736 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96cn9" event={"ID":"cde9313c-9316-40dc-8427-a8683ce67b5b","Type":"ContainerDied","Data":"76a47db264d65b885bec218b6f08f2c697e7ce16475c28980a69fb733df1f546"} Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.411800 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96cn9" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.416701 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-555rd" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.422186 4808 scope.go:117] "RemoveContainer" containerID="06d164f15f4ffda32fc38e8dd0fc8362cd4ad5865f18862e868ad682cb37fbff" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.427882 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fwnlt" event={"ID":"f11d89a2-be43-47fd-ae09-57fcd7df9820","Type":"ContainerDied","Data":"58f125d1375deb8d27747d24dbe4796c293d2634ffffc55921e3d616f23e9877"} Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.427926 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fwnlt" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.433290 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" event={"ID":"94a5dac5-a893-481f-ab4e-bfbd4106b8c6","Type":"ContainerDied","Data":"3799a0fbac72e744e334d592b8c23c266390e3b8efe6eab592facaf0d60c4dda"} Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.433367 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2v2lb" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.450720 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-555rd" podStartSLOduration=1.450659803 podStartE2EDuration="1.450659803s" podCreationTimestamp="2025-11-24 17:31:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:31:07.43715414 +0000 UTC m=+260.034821942" watchObservedRunningTime="2025-11-24 17:31:07.450659803 +0000 UTC m=+260.048327615" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.457630 4808 scope.go:117] "RemoveContainer" containerID="64744792fad2fd91ce4c09fe70f2e2d13d7d9c14962852daddadaf8d1d791551" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.461099 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mlmvz"] Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.464691 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mlmvz"] Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.487480 4808 scope.go:117] "RemoveContainer" containerID="2d7a3cb59965a53140d64612d3ab007f76ea1382eeef72510e0d469dee1c791a" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.515661 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fwnlt"] Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.518128 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fwnlt"] Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.525749 4808 scope.go:117] "RemoveContainer" containerID="d243094397bae39dc1bcdc98fafeadb91f71740003cda8386cd069d03bc0b92c" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.526546 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5wdrd"] Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.529877 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5wdrd"] Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.538877 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2v2lb"] Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.543592 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2v2lb"] Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.560383 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-96cn9"] Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.560766 4808 scope.go:117] "RemoveContainer" containerID="e37a0207952e492c77dcd4b1371413b00d9a0080255c7b4dcf56ce80031d5e6f" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.564062 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-96cn9"] Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.576178 4808 scope.go:117] "RemoveContainer" containerID="69001147dd08c93f8adef7b397ecfe5c5cdd8992418d86e2a9624b691cdeee11" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.594075 4808 scope.go:117] "RemoveContainer" containerID="a20f267a55e4331b4c5d56204677052a620e963ed21bd48f5c83bd080d6e5e83" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.606752 4808 scope.go:117] "RemoveContainer" containerID="01542eaca128f2d4d04cb61fdc5f02e4cf7503fca6f0f0de6b39740c2863931b" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.619300 4808 scope.go:117] "RemoveContainer" containerID="232dc4691cf01a2e914d3b89c3027a795a276f026a1fd2558510f1e52e155a22" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.638689 4808 scope.go:117] "RemoveContainer" containerID="bbad2b38eeb4ef59ee7db3e9900cd9e50c77515e0f1b3ffd815c8b4c7cdefd99" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.659927 4808 scope.go:117] "RemoveContainer" containerID="898c3b72ff56e4f287688b51154b00b934a3e1b3f59e0ac1997e09e3160f18e8" Nov 24 17:31:07 crc kubenswrapper[4808]: I1124 17:31:07.676329 4808 scope.go:117] "RemoveContainer" containerID="8b1591ff5b64fd5ed0c2ae4b6e0e871c54201749042134594259363a5cf31f3e" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.335355 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pxdgf"] Nov 24 17:31:08 crc kubenswrapper[4808]: E1124 17:31:08.336062 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cde9313c-9316-40dc-8427-a8683ce67b5b" containerName="extract-content" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336077 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="cde9313c-9316-40dc-8427-a8683ce67b5b" containerName="extract-content" Nov 24 17:31:08 crc kubenswrapper[4808]: E1124 17:31:08.336086 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94a5dac5-a893-481f-ab4e-bfbd4106b8c6" containerName="marketplace-operator" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336093 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="94a5dac5-a893-481f-ab4e-bfbd4106b8c6" containerName="marketplace-operator" Nov 24 17:31:08 crc kubenswrapper[4808]: E1124 17:31:08.336106 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b005cb77-3bba-44a5-8446-f6fcc5053336" containerName="registry-server" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336113 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b005cb77-3bba-44a5-8446-f6fcc5053336" containerName="registry-server" Nov 24 17:31:08 crc kubenswrapper[4808]: E1124 17:31:08.336121 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3f2246d-a6bb-463f-9a38-2d7144079eb6" containerName="extract-utilities" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336127 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3f2246d-a6bb-463f-9a38-2d7144079eb6" containerName="extract-utilities" Nov 24 17:31:08 crc kubenswrapper[4808]: E1124 17:31:08.336137 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b005cb77-3bba-44a5-8446-f6fcc5053336" containerName="extract-utilities" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336144 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b005cb77-3bba-44a5-8446-f6fcc5053336" containerName="extract-utilities" Nov 24 17:31:08 crc kubenswrapper[4808]: E1124 17:31:08.336151 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f11d89a2-be43-47fd-ae09-57fcd7df9820" containerName="extract-utilities" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336158 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f11d89a2-be43-47fd-ae09-57fcd7df9820" containerName="extract-utilities" Nov 24 17:31:08 crc kubenswrapper[4808]: E1124 17:31:08.336167 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3f2246d-a6bb-463f-9a38-2d7144079eb6" containerName="registry-server" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336173 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3f2246d-a6bb-463f-9a38-2d7144079eb6" containerName="registry-server" Nov 24 17:31:08 crc kubenswrapper[4808]: E1124 17:31:08.336181 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cde9313c-9316-40dc-8427-a8683ce67b5b" containerName="registry-server" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336189 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="cde9313c-9316-40dc-8427-a8683ce67b5b" containerName="registry-server" Nov 24 17:31:08 crc kubenswrapper[4808]: E1124 17:31:08.336200 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3f2246d-a6bb-463f-9a38-2d7144079eb6" containerName="extract-content" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336209 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3f2246d-a6bb-463f-9a38-2d7144079eb6" containerName="extract-content" Nov 24 17:31:08 crc kubenswrapper[4808]: E1124 17:31:08.336217 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b005cb77-3bba-44a5-8446-f6fcc5053336" containerName="extract-content" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336223 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b005cb77-3bba-44a5-8446-f6fcc5053336" containerName="extract-content" Nov 24 17:31:08 crc kubenswrapper[4808]: E1124 17:31:08.336230 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f11d89a2-be43-47fd-ae09-57fcd7df9820" containerName="extract-content" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336237 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f11d89a2-be43-47fd-ae09-57fcd7df9820" containerName="extract-content" Nov 24 17:31:08 crc kubenswrapper[4808]: E1124 17:31:08.336249 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f11d89a2-be43-47fd-ae09-57fcd7df9820" containerName="registry-server" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336254 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f11d89a2-be43-47fd-ae09-57fcd7df9820" containerName="registry-server" Nov 24 17:31:08 crc kubenswrapper[4808]: E1124 17:31:08.336266 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cde9313c-9316-40dc-8427-a8683ce67b5b" containerName="extract-utilities" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336273 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="cde9313c-9316-40dc-8427-a8683ce67b5b" containerName="extract-utilities" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336364 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b005cb77-3bba-44a5-8446-f6fcc5053336" containerName="registry-server" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336375 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3f2246d-a6bb-463f-9a38-2d7144079eb6" containerName="registry-server" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336386 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f11d89a2-be43-47fd-ae09-57fcd7df9820" containerName="registry-server" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336394 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="94a5dac5-a893-481f-ab4e-bfbd4106b8c6" containerName="marketplace-operator" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.336404 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="cde9313c-9316-40dc-8427-a8683ce67b5b" containerName="registry-server" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.337701 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pxdgf" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.340527 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.360608 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94a5dac5-a893-481f-ab4e-bfbd4106b8c6" path="/var/lib/kubelet/pods/94a5dac5-a893-481f-ab4e-bfbd4106b8c6/volumes" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.366405 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3f2246d-a6bb-463f-9a38-2d7144079eb6" path="/var/lib/kubelet/pods/a3f2246d-a6bb-463f-9a38-2d7144079eb6/volumes" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.367338 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b005cb77-3bba-44a5-8446-f6fcc5053336" path="/var/lib/kubelet/pods/b005cb77-3bba-44a5-8446-f6fcc5053336/volumes" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.368812 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cde9313c-9316-40dc-8427-a8683ce67b5b" path="/var/lib/kubelet/pods/cde9313c-9316-40dc-8427-a8683ce67b5b/volumes" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.369709 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f11d89a2-be43-47fd-ae09-57fcd7df9820" path="/var/lib/kubelet/pods/f11d89a2-be43-47fd-ae09-57fcd7df9820/volumes" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.370523 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pxdgf"] Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.437747 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14cc0e85-6bcb-443d-80f9-5f142532656a-utilities\") pod \"redhat-marketplace-pxdgf\" (UID: \"14cc0e85-6bcb-443d-80f9-5f142532656a\") " pod="openshift-marketplace/redhat-marketplace-pxdgf" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.437791 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14cc0e85-6bcb-443d-80f9-5f142532656a-catalog-content\") pod \"redhat-marketplace-pxdgf\" (UID: \"14cc0e85-6bcb-443d-80f9-5f142532656a\") " pod="openshift-marketplace/redhat-marketplace-pxdgf" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.437831 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwxww\" (UniqueName: \"kubernetes.io/projected/14cc0e85-6bcb-443d-80f9-5f142532656a-kube-api-access-fwxww\") pod \"redhat-marketplace-pxdgf\" (UID: \"14cc0e85-6bcb-443d-80f9-5f142532656a\") " pod="openshift-marketplace/redhat-marketplace-pxdgf" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.529098 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cnl6r"] Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.530976 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.536208 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.538997 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwxww\" (UniqueName: \"kubernetes.io/projected/14cc0e85-6bcb-443d-80f9-5f142532656a-kube-api-access-fwxww\") pod \"redhat-marketplace-pxdgf\" (UID: \"14cc0e85-6bcb-443d-80f9-5f142532656a\") " pod="openshift-marketplace/redhat-marketplace-pxdgf" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.541026 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cnl6r"] Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.541869 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14cc0e85-6bcb-443d-80f9-5f142532656a-utilities\") pod \"redhat-marketplace-pxdgf\" (UID: \"14cc0e85-6bcb-443d-80f9-5f142532656a\") " pod="openshift-marketplace/redhat-marketplace-pxdgf" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.541907 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14cc0e85-6bcb-443d-80f9-5f142532656a-catalog-content\") pod \"redhat-marketplace-pxdgf\" (UID: \"14cc0e85-6bcb-443d-80f9-5f142532656a\") " pod="openshift-marketplace/redhat-marketplace-pxdgf" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.545753 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14cc0e85-6bcb-443d-80f9-5f142532656a-catalog-content\") pod \"redhat-marketplace-pxdgf\" (UID: \"14cc0e85-6bcb-443d-80f9-5f142532656a\") " pod="openshift-marketplace/redhat-marketplace-pxdgf" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.548778 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14cc0e85-6bcb-443d-80f9-5f142532656a-utilities\") pod \"redhat-marketplace-pxdgf\" (UID: \"14cc0e85-6bcb-443d-80f9-5f142532656a\") " pod="openshift-marketplace/redhat-marketplace-pxdgf" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.573035 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwxww\" (UniqueName: \"kubernetes.io/projected/14cc0e85-6bcb-443d-80f9-5f142532656a-kube-api-access-fwxww\") pod \"redhat-marketplace-pxdgf\" (UID: \"14cc0e85-6bcb-443d-80f9-5f142532656a\") " pod="openshift-marketplace/redhat-marketplace-pxdgf" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.643353 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c935237c-9ccc-49e7-81b8-c770dcc3665b-utilities\") pod \"redhat-operators-cnl6r\" (UID: \"c935237c-9ccc-49e7-81b8-c770dcc3665b\") " pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.643420 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c935237c-9ccc-49e7-81b8-c770dcc3665b-catalog-content\") pod \"redhat-operators-cnl6r\" (UID: \"c935237c-9ccc-49e7-81b8-c770dcc3665b\") " pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.643490 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjlk9\" (UniqueName: \"kubernetes.io/projected/c935237c-9ccc-49e7-81b8-c770dcc3665b-kube-api-access-pjlk9\") pod \"redhat-operators-cnl6r\" (UID: \"c935237c-9ccc-49e7-81b8-c770dcc3665b\") " pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.653325 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pxdgf" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.744530 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjlk9\" (UniqueName: \"kubernetes.io/projected/c935237c-9ccc-49e7-81b8-c770dcc3665b-kube-api-access-pjlk9\") pod \"redhat-operators-cnl6r\" (UID: \"c935237c-9ccc-49e7-81b8-c770dcc3665b\") " pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.745422 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c935237c-9ccc-49e7-81b8-c770dcc3665b-utilities\") pod \"redhat-operators-cnl6r\" (UID: \"c935237c-9ccc-49e7-81b8-c770dcc3665b\") " pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.745468 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c935237c-9ccc-49e7-81b8-c770dcc3665b-catalog-content\") pod \"redhat-operators-cnl6r\" (UID: \"c935237c-9ccc-49e7-81b8-c770dcc3665b\") " pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.746058 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c935237c-9ccc-49e7-81b8-c770dcc3665b-catalog-content\") pod \"redhat-operators-cnl6r\" (UID: \"c935237c-9ccc-49e7-81b8-c770dcc3665b\") " pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.746563 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c935237c-9ccc-49e7-81b8-c770dcc3665b-utilities\") pod \"redhat-operators-cnl6r\" (UID: \"c935237c-9ccc-49e7-81b8-c770dcc3665b\") " pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.765976 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjlk9\" (UniqueName: \"kubernetes.io/projected/c935237c-9ccc-49e7-81b8-c770dcc3665b-kube-api-access-pjlk9\") pod \"redhat-operators-cnl6r\" (UID: \"c935237c-9ccc-49e7-81b8-c770dcc3665b\") " pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.850700 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pxdgf"] Nov 24 17:31:08 crc kubenswrapper[4808]: I1124 17:31:08.852542 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 17:31:09 crc kubenswrapper[4808]: I1124 17:31:09.070961 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cnl6r"] Nov 24 17:31:09 crc kubenswrapper[4808]: W1124 17:31:09.075184 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc935237c_9ccc_49e7_81b8_c770dcc3665b.slice/crio-6d5cd38e4f0ad115fe9191b1ea908f7821c876dd5957e88890e0f31813c3ecd1 WatchSource:0}: Error finding container 6d5cd38e4f0ad115fe9191b1ea908f7821c876dd5957e88890e0f31813c3ecd1: Status 404 returned error can't find the container with id 6d5cd38e4f0ad115fe9191b1ea908f7821c876dd5957e88890e0f31813c3ecd1 Nov 24 17:31:09 crc kubenswrapper[4808]: I1124 17:31:09.473910 4808 generic.go:334] "Generic (PLEG): container finished" podID="14cc0e85-6bcb-443d-80f9-5f142532656a" containerID="61ed971b3b066f32d3241993d3489bff74ff866218b89816d9ad3819ee8df2f6" exitCode=0 Nov 24 17:31:09 crc kubenswrapper[4808]: I1124 17:31:09.473989 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pxdgf" event={"ID":"14cc0e85-6bcb-443d-80f9-5f142532656a","Type":"ContainerDied","Data":"61ed971b3b066f32d3241993d3489bff74ff866218b89816d9ad3819ee8df2f6"} Nov 24 17:31:09 crc kubenswrapper[4808]: I1124 17:31:09.474037 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pxdgf" event={"ID":"14cc0e85-6bcb-443d-80f9-5f142532656a","Type":"ContainerStarted","Data":"55255d699ad2667ae889fb4947592a6a9a85c4f6f161a4f128d7913d2bd7d2ea"} Nov 24 17:31:09 crc kubenswrapper[4808]: I1124 17:31:09.480543 4808 generic.go:334] "Generic (PLEG): container finished" podID="c935237c-9ccc-49e7-81b8-c770dcc3665b" containerID="eb216262205a74dafc233cd764d420cbfac697565bfdded8259afc744dad1608" exitCode=0 Nov 24 17:31:09 crc kubenswrapper[4808]: I1124 17:31:09.481529 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cnl6r" event={"ID":"c935237c-9ccc-49e7-81b8-c770dcc3665b","Type":"ContainerDied","Data":"eb216262205a74dafc233cd764d420cbfac697565bfdded8259afc744dad1608"} Nov 24 17:31:09 crc kubenswrapper[4808]: I1124 17:31:09.481574 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cnl6r" event={"ID":"c935237c-9ccc-49e7-81b8-c770dcc3665b","Type":"ContainerStarted","Data":"6d5cd38e4f0ad115fe9191b1ea908f7821c876dd5957e88890e0f31813c3ecd1"} Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.730330 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5j6bd"] Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.731984 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5j6bd" Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.734546 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.742969 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5j6bd"] Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.878045 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf7q5\" (UniqueName: \"kubernetes.io/projected/ba197cbd-9042-410f-adb4-d05cfec94ead-kube-api-access-wf7q5\") pod \"certified-operators-5j6bd\" (UID: \"ba197cbd-9042-410f-adb4-d05cfec94ead\") " pod="openshift-marketplace/certified-operators-5j6bd" Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.878104 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba197cbd-9042-410f-adb4-d05cfec94ead-catalog-content\") pod \"certified-operators-5j6bd\" (UID: \"ba197cbd-9042-410f-adb4-d05cfec94ead\") " pod="openshift-marketplace/certified-operators-5j6bd" Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.878132 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba197cbd-9042-410f-adb4-d05cfec94ead-utilities\") pod \"certified-operators-5j6bd\" (UID: \"ba197cbd-9042-410f-adb4-d05cfec94ead\") " pod="openshift-marketplace/certified-operators-5j6bd" Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.936325 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rbql2"] Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.942916 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rbql2" Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.951437 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.967650 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rbql2"] Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.979554 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf7q5\" (UniqueName: \"kubernetes.io/projected/ba197cbd-9042-410f-adb4-d05cfec94ead-kube-api-access-wf7q5\") pod \"certified-operators-5j6bd\" (UID: \"ba197cbd-9042-410f-adb4-d05cfec94ead\") " pod="openshift-marketplace/certified-operators-5j6bd" Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.979625 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba197cbd-9042-410f-adb4-d05cfec94ead-catalog-content\") pod \"certified-operators-5j6bd\" (UID: \"ba197cbd-9042-410f-adb4-d05cfec94ead\") " pod="openshift-marketplace/certified-operators-5j6bd" Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.979650 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba197cbd-9042-410f-adb4-d05cfec94ead-utilities\") pod \"certified-operators-5j6bd\" (UID: \"ba197cbd-9042-410f-adb4-d05cfec94ead\") " pod="openshift-marketplace/certified-operators-5j6bd" Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.980316 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba197cbd-9042-410f-adb4-d05cfec94ead-utilities\") pod \"certified-operators-5j6bd\" (UID: \"ba197cbd-9042-410f-adb4-d05cfec94ead\") " pod="openshift-marketplace/certified-operators-5j6bd" Nov 24 17:31:10 crc kubenswrapper[4808]: I1124 17:31:10.980528 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba197cbd-9042-410f-adb4-d05cfec94ead-catalog-content\") pod \"certified-operators-5j6bd\" (UID: \"ba197cbd-9042-410f-adb4-d05cfec94ead\") " pod="openshift-marketplace/certified-operators-5j6bd" Nov 24 17:31:11 crc kubenswrapper[4808]: I1124 17:31:11.008140 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf7q5\" (UniqueName: \"kubernetes.io/projected/ba197cbd-9042-410f-adb4-d05cfec94ead-kube-api-access-wf7q5\") pod \"certified-operators-5j6bd\" (UID: \"ba197cbd-9042-410f-adb4-d05cfec94ead\") " pod="openshift-marketplace/certified-operators-5j6bd" Nov 24 17:31:11 crc kubenswrapper[4808]: I1124 17:31:11.079611 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5j6bd" Nov 24 17:31:11 crc kubenswrapper[4808]: I1124 17:31:11.081627 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7929b2f8-b7af-4a85-9150-861fe438f602-utilities\") pod \"community-operators-rbql2\" (UID: \"7929b2f8-b7af-4a85-9150-861fe438f602\") " pod="openshift-marketplace/community-operators-rbql2" Nov 24 17:31:11 crc kubenswrapper[4808]: I1124 17:31:11.081900 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsbcb\" (UniqueName: \"kubernetes.io/projected/7929b2f8-b7af-4a85-9150-861fe438f602-kube-api-access-tsbcb\") pod \"community-operators-rbql2\" (UID: \"7929b2f8-b7af-4a85-9150-861fe438f602\") " pod="openshift-marketplace/community-operators-rbql2" Nov 24 17:31:11 crc kubenswrapper[4808]: I1124 17:31:11.083853 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7929b2f8-b7af-4a85-9150-861fe438f602-catalog-content\") pod \"community-operators-rbql2\" (UID: \"7929b2f8-b7af-4a85-9150-861fe438f602\") " pod="openshift-marketplace/community-operators-rbql2" Nov 24 17:31:11 crc kubenswrapper[4808]: I1124 17:31:11.186325 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7929b2f8-b7af-4a85-9150-861fe438f602-utilities\") pod \"community-operators-rbql2\" (UID: \"7929b2f8-b7af-4a85-9150-861fe438f602\") " pod="openshift-marketplace/community-operators-rbql2" Nov 24 17:31:11 crc kubenswrapper[4808]: I1124 17:31:11.186414 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsbcb\" (UniqueName: \"kubernetes.io/projected/7929b2f8-b7af-4a85-9150-861fe438f602-kube-api-access-tsbcb\") pod \"community-operators-rbql2\" (UID: \"7929b2f8-b7af-4a85-9150-861fe438f602\") " pod="openshift-marketplace/community-operators-rbql2" Nov 24 17:31:11 crc kubenswrapper[4808]: I1124 17:31:11.186498 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7929b2f8-b7af-4a85-9150-861fe438f602-catalog-content\") pod \"community-operators-rbql2\" (UID: \"7929b2f8-b7af-4a85-9150-861fe438f602\") " pod="openshift-marketplace/community-operators-rbql2" Nov 24 17:31:11 crc kubenswrapper[4808]: I1124 17:31:11.187261 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7929b2f8-b7af-4a85-9150-861fe438f602-catalog-content\") pod \"community-operators-rbql2\" (UID: \"7929b2f8-b7af-4a85-9150-861fe438f602\") " pod="openshift-marketplace/community-operators-rbql2" Nov 24 17:31:11 crc kubenswrapper[4808]: I1124 17:31:11.187613 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7929b2f8-b7af-4a85-9150-861fe438f602-utilities\") pod \"community-operators-rbql2\" (UID: \"7929b2f8-b7af-4a85-9150-861fe438f602\") " pod="openshift-marketplace/community-operators-rbql2" Nov 24 17:31:11 crc kubenswrapper[4808]: I1124 17:31:11.209328 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsbcb\" (UniqueName: \"kubernetes.io/projected/7929b2f8-b7af-4a85-9150-861fe438f602-kube-api-access-tsbcb\") pod \"community-operators-rbql2\" (UID: \"7929b2f8-b7af-4a85-9150-861fe438f602\") " pod="openshift-marketplace/community-operators-rbql2" Nov 24 17:31:11 crc kubenswrapper[4808]: I1124 17:31:11.273378 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rbql2" Nov 24 17:31:11 crc kubenswrapper[4808]: I1124 17:31:11.319165 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5j6bd"] Nov 24 17:31:11 crc kubenswrapper[4808]: W1124 17:31:11.328676 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba197cbd_9042_410f_adb4_d05cfec94ead.slice/crio-da4d622e544828f8fd83f82aae1400ff204c8a4a0682b9629c5329ea18e4ca05 WatchSource:0}: Error finding container da4d622e544828f8fd83f82aae1400ff204c8a4a0682b9629c5329ea18e4ca05: Status 404 returned error can't find the container with id da4d622e544828f8fd83f82aae1400ff204c8a4a0682b9629c5329ea18e4ca05 Nov 24 17:31:11 crc kubenswrapper[4808]: I1124 17:31:11.491351 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j6bd" event={"ID":"ba197cbd-9042-410f-adb4-d05cfec94ead","Type":"ContainerStarted","Data":"da4d622e544828f8fd83f82aae1400ff204c8a4a0682b9629c5329ea18e4ca05"} Nov 24 17:31:12 crc kubenswrapper[4808]: I1124 17:31:12.024875 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rbql2"] Nov 24 17:31:12 crc kubenswrapper[4808]: W1124 17:31:12.036886 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7929b2f8_b7af_4a85_9150_861fe438f602.slice/crio-a2de7658ff6fe555f774d300b0f094065b32714e4180958ca6d8fd330d3cc653 WatchSource:0}: Error finding container a2de7658ff6fe555f774d300b0f094065b32714e4180958ca6d8fd330d3cc653: Status 404 returned error can't find the container with id a2de7658ff6fe555f774d300b0f094065b32714e4180958ca6d8fd330d3cc653 Nov 24 17:31:12 crc kubenswrapper[4808]: I1124 17:31:12.499098 4808 generic.go:334] "Generic (PLEG): container finished" podID="ba197cbd-9042-410f-adb4-d05cfec94ead" containerID="36d83b5f0b05c768d2f34468c4e115694f9ce4fd37f689f1d676267a20ffdfa4" exitCode=0 Nov 24 17:31:12 crc kubenswrapper[4808]: I1124 17:31:12.499198 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j6bd" event={"ID":"ba197cbd-9042-410f-adb4-d05cfec94ead","Type":"ContainerDied","Data":"36d83b5f0b05c768d2f34468c4e115694f9ce4fd37f689f1d676267a20ffdfa4"} Nov 24 17:31:12 crc kubenswrapper[4808]: I1124 17:31:12.504319 4808 generic.go:334] "Generic (PLEG): container finished" podID="7929b2f8-b7af-4a85-9150-861fe438f602" containerID="0b7c95c5b85ab0b33094da96f0ac554d32c34cdea1e05eb13961e7ffe2342592" exitCode=0 Nov 24 17:31:12 crc kubenswrapper[4808]: I1124 17:31:12.504431 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rbql2" event={"ID":"7929b2f8-b7af-4a85-9150-861fe438f602","Type":"ContainerDied","Data":"0b7c95c5b85ab0b33094da96f0ac554d32c34cdea1e05eb13961e7ffe2342592"} Nov 24 17:31:12 crc kubenswrapper[4808]: I1124 17:31:12.504514 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rbql2" event={"ID":"7929b2f8-b7af-4a85-9150-861fe438f602","Type":"ContainerStarted","Data":"a2de7658ff6fe555f774d300b0f094065b32714e4180958ca6d8fd330d3cc653"} Nov 24 17:31:12 crc kubenswrapper[4808]: I1124 17:31:12.509674 4808 generic.go:334] "Generic (PLEG): container finished" podID="14cc0e85-6bcb-443d-80f9-5f142532656a" containerID="1ef2aa60e6f8f0b84487c3f794409220e67ab7ed6cca3404869a5ea009a7457b" exitCode=0 Nov 24 17:31:12 crc kubenswrapper[4808]: I1124 17:31:12.509772 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pxdgf" event={"ID":"14cc0e85-6bcb-443d-80f9-5f142532656a","Type":"ContainerDied","Data":"1ef2aa60e6f8f0b84487c3f794409220e67ab7ed6cca3404869a5ea009a7457b"} Nov 24 17:31:12 crc kubenswrapper[4808]: I1124 17:31:12.514243 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cnl6r" event={"ID":"c935237c-9ccc-49e7-81b8-c770dcc3665b","Type":"ContainerStarted","Data":"582d3ed297d4554f523e547b529107bd21accfd72fe0c71ce521531b0c7210ca"} Nov 24 17:31:13 crc kubenswrapper[4808]: I1124 17:31:13.523991 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pxdgf" event={"ID":"14cc0e85-6bcb-443d-80f9-5f142532656a","Type":"ContainerStarted","Data":"5f60b766cedccda6054ec3527d4f33e2805199d8e756670a99be3b9fefecb85e"} Nov 24 17:31:13 crc kubenswrapper[4808]: I1124 17:31:13.526456 4808 generic.go:334] "Generic (PLEG): container finished" podID="c935237c-9ccc-49e7-81b8-c770dcc3665b" containerID="582d3ed297d4554f523e547b529107bd21accfd72fe0c71ce521531b0c7210ca" exitCode=0 Nov 24 17:31:13 crc kubenswrapper[4808]: I1124 17:31:13.526537 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cnl6r" event={"ID":"c935237c-9ccc-49e7-81b8-c770dcc3665b","Type":"ContainerDied","Data":"582d3ed297d4554f523e547b529107bd21accfd72fe0c71ce521531b0c7210ca"} Nov 24 17:31:13 crc kubenswrapper[4808]: I1124 17:31:13.529612 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j6bd" event={"ID":"ba197cbd-9042-410f-adb4-d05cfec94ead","Type":"ContainerStarted","Data":"ba5ec282e414f50c853902acb67efe1fe5527924d2b5c6aea738b27c10c956f8"} Nov 24 17:31:13 crc kubenswrapper[4808]: I1124 17:31:13.533488 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rbql2" event={"ID":"7929b2f8-b7af-4a85-9150-861fe438f602","Type":"ContainerStarted","Data":"9ba136e89a344dcd19f8ae4ec641546a84d6dc854f5ea63123c2e805e59d49e7"} Nov 24 17:31:13 crc kubenswrapper[4808]: I1124 17:31:13.549511 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pxdgf" podStartSLOduration=2.096318166 podStartE2EDuration="5.549483401s" podCreationTimestamp="2025-11-24 17:31:08 +0000 UTC" firstStartedPulling="2025-11-24 17:31:09.476516916 +0000 UTC m=+262.074184718" lastFinishedPulling="2025-11-24 17:31:12.929682151 +0000 UTC m=+265.527349953" observedRunningTime="2025-11-24 17:31:13.54912918 +0000 UTC m=+266.146796992" watchObservedRunningTime="2025-11-24 17:31:13.549483401 +0000 UTC m=+266.147151213" Nov 24 17:31:14 crc kubenswrapper[4808]: I1124 17:31:14.541036 4808 generic.go:334] "Generic (PLEG): container finished" podID="ba197cbd-9042-410f-adb4-d05cfec94ead" containerID="ba5ec282e414f50c853902acb67efe1fe5527924d2b5c6aea738b27c10c956f8" exitCode=0 Nov 24 17:31:14 crc kubenswrapper[4808]: I1124 17:31:14.541128 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j6bd" event={"ID":"ba197cbd-9042-410f-adb4-d05cfec94ead","Type":"ContainerDied","Data":"ba5ec282e414f50c853902acb67efe1fe5527924d2b5c6aea738b27c10c956f8"} Nov 24 17:31:14 crc kubenswrapper[4808]: I1124 17:31:14.547329 4808 generic.go:334] "Generic (PLEG): container finished" podID="7929b2f8-b7af-4a85-9150-861fe438f602" containerID="9ba136e89a344dcd19f8ae4ec641546a84d6dc854f5ea63123c2e805e59d49e7" exitCode=0 Nov 24 17:31:14 crc kubenswrapper[4808]: I1124 17:31:14.547433 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rbql2" event={"ID":"7929b2f8-b7af-4a85-9150-861fe438f602","Type":"ContainerDied","Data":"9ba136e89a344dcd19f8ae4ec641546a84d6dc854f5ea63123c2e805e59d49e7"} Nov 24 17:31:14 crc kubenswrapper[4808]: I1124 17:31:14.547470 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rbql2" event={"ID":"7929b2f8-b7af-4a85-9150-861fe438f602","Type":"ContainerStarted","Data":"addd2109e0933d1bed83754b2bb4cb3e3bf97af18368d6f6d4b14fd465fb62d7"} Nov 24 17:31:14 crc kubenswrapper[4808]: I1124 17:31:14.555687 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cnl6r" event={"ID":"c935237c-9ccc-49e7-81b8-c770dcc3665b","Type":"ContainerStarted","Data":"f596ec07b1f76054fd189d12901fca0476be5d8f6f131e5740c03581cbeeba8d"} Nov 24 17:31:14 crc kubenswrapper[4808]: I1124 17:31:14.644695 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rbql2" podStartSLOduration=3.176135152 podStartE2EDuration="4.644673156s" podCreationTimestamp="2025-11-24 17:31:10 +0000 UTC" firstStartedPulling="2025-11-24 17:31:12.50594741 +0000 UTC m=+265.103615212" lastFinishedPulling="2025-11-24 17:31:13.974485424 +0000 UTC m=+266.572153216" observedRunningTime="2025-11-24 17:31:14.641982367 +0000 UTC m=+267.239650189" watchObservedRunningTime="2025-11-24 17:31:14.644673156 +0000 UTC m=+267.242340958" Nov 24 17:31:14 crc kubenswrapper[4808]: I1124 17:31:14.644983 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cnl6r" podStartSLOduration=2.235378054 podStartE2EDuration="6.644978526s" podCreationTimestamp="2025-11-24 17:31:08 +0000 UTC" firstStartedPulling="2025-11-24 17:31:09.482113021 +0000 UTC m=+262.079780833" lastFinishedPulling="2025-11-24 17:31:13.891713513 +0000 UTC m=+266.489381305" observedRunningTime="2025-11-24 17:31:14.625880486 +0000 UTC m=+267.223548288" watchObservedRunningTime="2025-11-24 17:31:14.644978526 +0000 UTC m=+267.242646318" Nov 24 17:31:16 crc kubenswrapper[4808]: I1124 17:31:16.566842 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j6bd" event={"ID":"ba197cbd-9042-410f-adb4-d05cfec94ead","Type":"ContainerStarted","Data":"725d7e7ce2ff3e08bf39148f44bde7614e4f1848406e31d7c23e26e4b7a62a99"} Nov 24 17:31:16 crc kubenswrapper[4808]: I1124 17:31:16.591165 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5j6bd" podStartSLOduration=4.1419565 podStartE2EDuration="6.591141539s" podCreationTimestamp="2025-11-24 17:31:10 +0000 UTC" firstStartedPulling="2025-11-24 17:31:12.504616656 +0000 UTC m=+265.102284458" lastFinishedPulling="2025-11-24 17:31:14.953801665 +0000 UTC m=+267.551469497" observedRunningTime="2025-11-24 17:31:16.586709633 +0000 UTC m=+269.184377435" watchObservedRunningTime="2025-11-24 17:31:16.591141539 +0000 UTC m=+269.188809341" Nov 24 17:31:18 crc kubenswrapper[4808]: I1124 17:31:18.654609 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pxdgf" Nov 24 17:31:18 crc kubenswrapper[4808]: I1124 17:31:18.655185 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pxdgf" Nov 24 17:31:18 crc kubenswrapper[4808]: I1124 17:31:18.707041 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pxdgf" Nov 24 17:31:18 crc kubenswrapper[4808]: I1124 17:31:18.853662 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 17:31:18 crc kubenswrapper[4808]: I1124 17:31:18.853785 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 17:31:19 crc kubenswrapper[4808]: I1124 17:31:19.631227 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pxdgf" Nov 24 17:31:19 crc kubenswrapper[4808]: I1124 17:31:19.892624 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cnl6r" podUID="c935237c-9ccc-49e7-81b8-c770dcc3665b" containerName="registry-server" probeResult="failure" output=< Nov 24 17:31:19 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Nov 24 17:31:19 crc kubenswrapper[4808]: > Nov 24 17:31:21 crc kubenswrapper[4808]: I1124 17:31:21.080184 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5j6bd" Nov 24 17:31:21 crc kubenswrapper[4808]: I1124 17:31:21.080589 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5j6bd" Nov 24 17:31:21 crc kubenswrapper[4808]: I1124 17:31:21.147792 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5j6bd" Nov 24 17:31:21 crc kubenswrapper[4808]: I1124 17:31:21.273816 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rbql2" Nov 24 17:31:21 crc kubenswrapper[4808]: I1124 17:31:21.273893 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rbql2" Nov 24 17:31:21 crc kubenswrapper[4808]: I1124 17:31:21.347541 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rbql2" Nov 24 17:31:21 crc kubenswrapper[4808]: I1124 17:31:21.640439 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rbql2" Nov 24 17:31:21 crc kubenswrapper[4808]: I1124 17:31:21.643057 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5j6bd" Nov 24 17:31:28 crc kubenswrapper[4808]: I1124 17:31:28.899352 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 17:31:28 crc kubenswrapper[4808]: I1124 17:31:28.950653 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 17:33:06 crc kubenswrapper[4808]: I1124 17:33:06.522907 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:33:06 crc kubenswrapper[4808]: I1124 17:33:06.523592 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.493189 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mdpx2"] Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.495601 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.513754 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mdpx2"] Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.523119 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.523334 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.558440 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/23752d4b-1823-4f39-967e-69ff257b6835-trusted-ca\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.558528 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/23752d4b-1823-4f39-967e-69ff257b6835-bound-sa-token\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.558552 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpch8\" (UniqueName: \"kubernetes.io/projected/23752d4b-1823-4f39-967e-69ff257b6835-kube-api-access-gpch8\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.558573 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/23752d4b-1823-4f39-967e-69ff257b6835-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.558606 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.558685 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/23752d4b-1823-4f39-967e-69ff257b6835-registry-tls\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.558715 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/23752d4b-1823-4f39-967e-69ff257b6835-registry-certificates\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.558751 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/23752d4b-1823-4f39-967e-69ff257b6835-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.579935 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.660296 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/23752d4b-1823-4f39-967e-69ff257b6835-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.660577 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/23752d4b-1823-4f39-967e-69ff257b6835-bound-sa-token\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.660650 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpch8\" (UniqueName: \"kubernetes.io/projected/23752d4b-1823-4f39-967e-69ff257b6835-kube-api-access-gpch8\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.660734 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/23752d4b-1823-4f39-967e-69ff257b6835-registry-tls\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.660809 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/23752d4b-1823-4f39-967e-69ff257b6835-registry-certificates\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.660888 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/23752d4b-1823-4f39-967e-69ff257b6835-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.660960 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/23752d4b-1823-4f39-967e-69ff257b6835-trusted-ca\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.661933 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/23752d4b-1823-4f39-967e-69ff257b6835-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.662454 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/23752d4b-1823-4f39-967e-69ff257b6835-trusted-ca\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.662671 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/23752d4b-1823-4f39-967e-69ff257b6835-registry-certificates\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.667289 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/23752d4b-1823-4f39-967e-69ff257b6835-registry-tls\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.667332 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/23752d4b-1823-4f39-967e-69ff257b6835-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.677728 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/23752d4b-1823-4f39-967e-69ff257b6835-bound-sa-token\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.677784 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpch8\" (UniqueName: \"kubernetes.io/projected/23752d4b-1823-4f39-967e-69ff257b6835-kube-api-access-gpch8\") pod \"image-registry-66df7c8f76-mdpx2\" (UID: \"23752d4b-1823-4f39-967e-69ff257b6835\") " pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.814902 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:36 crc kubenswrapper[4808]: I1124 17:33:36.993854 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mdpx2"] Nov 24 17:33:37 crc kubenswrapper[4808]: I1124 17:33:37.548100 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" event={"ID":"23752d4b-1823-4f39-967e-69ff257b6835","Type":"ContainerStarted","Data":"ce0c00e93987fdf8fc610a7dba93274763d90128951e0511c61dcd4f2fcf6ae9"} Nov 24 17:33:37 crc kubenswrapper[4808]: I1124 17:33:37.548484 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" event={"ID":"23752d4b-1823-4f39-967e-69ff257b6835","Type":"ContainerStarted","Data":"e7410cde8535b79721ac7ce96f9670ab5c777447c18f27eb09a5e9c2d90d696e"} Nov 24 17:33:37 crc kubenswrapper[4808]: I1124 17:33:37.548625 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:37 crc kubenswrapper[4808]: I1124 17:33:37.577458 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" podStartSLOduration=1.5774381370000001 podStartE2EDuration="1.577438137s" podCreationTimestamp="2025-11-24 17:33:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:33:37.569184307 +0000 UTC m=+410.166852119" watchObservedRunningTime="2025-11-24 17:33:37.577438137 +0000 UTC m=+410.175105939" Nov 24 17:33:56 crc kubenswrapper[4808]: I1124 17:33:56.821001 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-mdpx2" Nov 24 17:33:56 crc kubenswrapper[4808]: I1124 17:33:56.915318 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gjnx8"] Nov 24 17:34:06 crc kubenswrapper[4808]: I1124 17:34:06.522588 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:34:06 crc kubenswrapper[4808]: I1124 17:34:06.523160 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:34:06 crc kubenswrapper[4808]: I1124 17:34:06.523252 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:34:06 crc kubenswrapper[4808]: I1124 17:34:06.523910 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4b9523f5392391855bf06e84ed62910c36d9c20fb3584ca98f4425fec2b9f41c"} pod="openshift-machine-config-operator/machine-config-daemon-ht87w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:34:06 crc kubenswrapper[4808]: I1124 17:34:06.523957 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" containerID="cri-o://4b9523f5392391855bf06e84ed62910c36d9c20fb3584ca98f4425fec2b9f41c" gracePeriod=600 Nov 24 17:34:07 crc kubenswrapper[4808]: I1124 17:34:07.722119 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerID="4b9523f5392391855bf06e84ed62910c36d9c20fb3584ca98f4425fec2b9f41c" exitCode=0 Nov 24 17:34:07 crc kubenswrapper[4808]: I1124 17:34:07.722194 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerDied","Data":"4b9523f5392391855bf06e84ed62910c36d9c20fb3584ca98f4425fec2b9f41c"} Nov 24 17:34:07 crc kubenswrapper[4808]: I1124 17:34:07.722255 4808 scope.go:117] "RemoveContainer" containerID="bb74c3a30d8bcb29348c16b174972de218b358ec2f80595db93c6464e719fca8" Nov 24 17:34:08 crc kubenswrapper[4808]: I1124 17:34:08.730480 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"8b5c11a71abbce839aaa56f6f8510529f68981025ad2c3fda226177d751485f2"} Nov 24 17:34:21 crc kubenswrapper[4808]: I1124 17:34:21.963069 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" podUID="edb7cea2-f6eb-4043-91d9-c7abaef74455" containerName="registry" containerID="cri-o://f4f10a51abdab5898fea49b8f4bafb394e8a9b487712f1bf33cf4cf93623b538" gracePeriod=30 Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.306376 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.386272 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/edb7cea2-f6eb-4043-91d9-c7abaef74455-trusted-ca\") pod \"edb7cea2-f6eb-4043-91d9-c7abaef74455\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.386339 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/edb7cea2-f6eb-4043-91d9-c7abaef74455-installation-pull-secrets\") pod \"edb7cea2-f6eb-4043-91d9-c7abaef74455\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.386363 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/edb7cea2-f6eb-4043-91d9-c7abaef74455-registry-certificates\") pod \"edb7cea2-f6eb-4043-91d9-c7abaef74455\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.386409 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-registry-tls\") pod \"edb7cea2-f6eb-4043-91d9-c7abaef74455\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.386458 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/edb7cea2-f6eb-4043-91d9-c7abaef74455-ca-trust-extracted\") pod \"edb7cea2-f6eb-4043-91d9-c7abaef74455\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.386532 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-bound-sa-token\") pod \"edb7cea2-f6eb-4043-91d9-c7abaef74455\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.386654 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqbq8\" (UniqueName: \"kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-kube-api-access-mqbq8\") pod \"edb7cea2-f6eb-4043-91d9-c7abaef74455\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.387280 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"edb7cea2-f6eb-4043-91d9-c7abaef74455\" (UID: \"edb7cea2-f6eb-4043-91d9-c7abaef74455\") " Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.389170 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edb7cea2-f6eb-4043-91d9-c7abaef74455-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "edb7cea2-f6eb-4043-91d9-c7abaef74455" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.389780 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edb7cea2-f6eb-4043-91d9-c7abaef74455-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "edb7cea2-f6eb-4043-91d9-c7abaef74455" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.395057 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edb7cea2-f6eb-4043-91d9-c7abaef74455-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "edb7cea2-f6eb-4043-91d9-c7abaef74455" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.395367 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "edb7cea2-f6eb-4043-91d9-c7abaef74455" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.395413 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "edb7cea2-f6eb-4043-91d9-c7abaef74455" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.398920 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-kube-api-access-mqbq8" (OuterVolumeSpecName: "kube-api-access-mqbq8") pod "edb7cea2-f6eb-4043-91d9-c7abaef74455" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455"). InnerVolumeSpecName "kube-api-access-mqbq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.400295 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "edb7cea2-f6eb-4043-91d9-c7abaef74455" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.408076 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edb7cea2-f6eb-4043-91d9-c7abaef74455-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "edb7cea2-f6eb-4043-91d9-c7abaef74455" (UID: "edb7cea2-f6eb-4043-91d9-c7abaef74455"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.489385 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/edb7cea2-f6eb-4043-91d9-c7abaef74455-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.489440 4808 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/edb7cea2-f6eb-4043-91d9-c7abaef74455-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.489453 4808 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/edb7cea2-f6eb-4043-91d9-c7abaef74455-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.489465 4808 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.489475 4808 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/edb7cea2-f6eb-4043-91d9-c7abaef74455-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.489484 4808 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.489494 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqbq8\" (UniqueName: \"kubernetes.io/projected/edb7cea2-f6eb-4043-91d9-c7abaef74455-kube-api-access-mqbq8\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.807870 4808 generic.go:334] "Generic (PLEG): container finished" podID="edb7cea2-f6eb-4043-91d9-c7abaef74455" containerID="f4f10a51abdab5898fea49b8f4bafb394e8a9b487712f1bf33cf4cf93623b538" exitCode=0 Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.807927 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" event={"ID":"edb7cea2-f6eb-4043-91d9-c7abaef74455","Type":"ContainerDied","Data":"f4f10a51abdab5898fea49b8f4bafb394e8a9b487712f1bf33cf4cf93623b538"} Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.807963 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.808298 4808 scope.go:117] "RemoveContainer" containerID="f4f10a51abdab5898fea49b8f4bafb394e8a9b487712f1bf33cf4cf93623b538" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.808281 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gjnx8" event={"ID":"edb7cea2-f6eb-4043-91d9-c7abaef74455","Type":"ContainerDied","Data":"2586c1c9c86057fce8c557ec2d913802200bd545faab365bf2fc119b93520556"} Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.826105 4808 scope.go:117] "RemoveContainer" containerID="f4f10a51abdab5898fea49b8f4bafb394e8a9b487712f1bf33cf4cf93623b538" Nov 24 17:34:22 crc kubenswrapper[4808]: E1124 17:34:22.826628 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4f10a51abdab5898fea49b8f4bafb394e8a9b487712f1bf33cf4cf93623b538\": container with ID starting with f4f10a51abdab5898fea49b8f4bafb394e8a9b487712f1bf33cf4cf93623b538 not found: ID does not exist" containerID="f4f10a51abdab5898fea49b8f4bafb394e8a9b487712f1bf33cf4cf93623b538" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.826663 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4f10a51abdab5898fea49b8f4bafb394e8a9b487712f1bf33cf4cf93623b538"} err="failed to get container status \"f4f10a51abdab5898fea49b8f4bafb394e8a9b487712f1bf33cf4cf93623b538\": rpc error: code = NotFound desc = could not find container \"f4f10a51abdab5898fea49b8f4bafb394e8a9b487712f1bf33cf4cf93623b538\": container with ID starting with f4f10a51abdab5898fea49b8f4bafb394e8a9b487712f1bf33cf4cf93623b538 not found: ID does not exist" Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.837620 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gjnx8"] Nov 24 17:34:22 crc kubenswrapper[4808]: I1124 17:34:22.841359 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gjnx8"] Nov 24 17:34:24 crc kubenswrapper[4808]: I1124 17:34:24.355905 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edb7cea2-f6eb-4043-91d9-c7abaef74455" path="/var/lib/kubelet/pods/edb7cea2-f6eb-4043-91d9-c7abaef74455/volumes" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.026494 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-cbqfz"] Nov 24 17:36:23 crc kubenswrapper[4808]: E1124 17:36:23.028490 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb7cea2-f6eb-4043-91d9-c7abaef74455" containerName="registry" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.028595 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb7cea2-f6eb-4043-91d9-c7abaef74455" containerName="registry" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.028821 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="edb7cea2-f6eb-4043-91d9-c7abaef74455" containerName="registry" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.029451 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-cbqfz" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.033326 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.038675 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.038906 4808 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-cbq7p" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.051110 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-cbqfz"] Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.060685 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-2qsx7"] Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.090845 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmf7g\" (UniqueName: \"kubernetes.io/projected/24eadecc-5782-4e5b-8982-46d3e99fa5b1-kube-api-access-vmf7g\") pod \"cert-manager-cainjector-7f985d654d-cbqfz\" (UID: \"24eadecc-5782-4e5b-8982-46d3e99fa5b1\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-cbqfz" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.094254 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-2qsx7" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.097322 4808 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-fxd66" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.103508 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-bjlpj"] Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.104639 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-bjlpj" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.106873 4808 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-49jgf" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.108062 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-bjlpj"] Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.110889 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-2qsx7"] Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.193141 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz2dg\" (UniqueName: \"kubernetes.io/projected/016a9f20-1203-4241-9210-55b50eeb60a1-kube-api-access-jz2dg\") pod \"cert-manager-5b446d88c5-2qsx7\" (UID: \"016a9f20-1203-4241-9210-55b50eeb60a1\") " pod="cert-manager/cert-manager-5b446d88c5-2qsx7" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.193266 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmf7g\" (UniqueName: \"kubernetes.io/projected/24eadecc-5782-4e5b-8982-46d3e99fa5b1-kube-api-access-vmf7g\") pod \"cert-manager-cainjector-7f985d654d-cbqfz\" (UID: \"24eadecc-5782-4e5b-8982-46d3e99fa5b1\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-cbqfz" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.193297 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfnxm\" (UniqueName: \"kubernetes.io/projected/a1ea24f0-1b28-4648-a09e-8c0539792ea1-kube-api-access-cfnxm\") pod \"cert-manager-webhook-5655c58dd6-bjlpj\" (UID: \"a1ea24f0-1b28-4648-a09e-8c0539792ea1\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-bjlpj" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.216998 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmf7g\" (UniqueName: \"kubernetes.io/projected/24eadecc-5782-4e5b-8982-46d3e99fa5b1-kube-api-access-vmf7g\") pod \"cert-manager-cainjector-7f985d654d-cbqfz\" (UID: \"24eadecc-5782-4e5b-8982-46d3e99fa5b1\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-cbqfz" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.294745 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz2dg\" (UniqueName: \"kubernetes.io/projected/016a9f20-1203-4241-9210-55b50eeb60a1-kube-api-access-jz2dg\") pod \"cert-manager-5b446d88c5-2qsx7\" (UID: \"016a9f20-1203-4241-9210-55b50eeb60a1\") " pod="cert-manager/cert-manager-5b446d88c5-2qsx7" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.295208 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfnxm\" (UniqueName: \"kubernetes.io/projected/a1ea24f0-1b28-4648-a09e-8c0539792ea1-kube-api-access-cfnxm\") pod \"cert-manager-webhook-5655c58dd6-bjlpj\" (UID: \"a1ea24f0-1b28-4648-a09e-8c0539792ea1\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-bjlpj" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.313508 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfnxm\" (UniqueName: \"kubernetes.io/projected/a1ea24f0-1b28-4648-a09e-8c0539792ea1-kube-api-access-cfnxm\") pod \"cert-manager-webhook-5655c58dd6-bjlpj\" (UID: \"a1ea24f0-1b28-4648-a09e-8c0539792ea1\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-bjlpj" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.315202 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz2dg\" (UniqueName: \"kubernetes.io/projected/016a9f20-1203-4241-9210-55b50eeb60a1-kube-api-access-jz2dg\") pod \"cert-manager-5b446d88c5-2qsx7\" (UID: \"016a9f20-1203-4241-9210-55b50eeb60a1\") " pod="cert-manager/cert-manager-5b446d88c5-2qsx7" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.352877 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-cbqfz" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.413459 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-2qsx7" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.421621 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-bjlpj" Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.668736 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-2qsx7"] Nov 24 17:36:23 crc kubenswrapper[4808]: W1124 17:36:23.673697 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod016a9f20_1203_4241_9210_55b50eeb60a1.slice/crio-b0291c771a93d83ded9407ab7e78c4f8c465796183ce18106baaefbe9bac4bb0 WatchSource:0}: Error finding container b0291c771a93d83ded9407ab7e78c4f8c465796183ce18106baaefbe9bac4bb0: Status 404 returned error can't find the container with id b0291c771a93d83ded9407ab7e78c4f8c465796183ce18106baaefbe9bac4bb0 Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.677150 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.718728 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-bjlpj"] Nov 24 17:36:23 crc kubenswrapper[4808]: W1124 17:36:23.727694 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1ea24f0_1b28_4648_a09e_8c0539792ea1.slice/crio-721354b0a01b76a735dbaabf45e2126b89c83fe8b16d4a79a6b303777d8bb264 WatchSource:0}: Error finding container 721354b0a01b76a735dbaabf45e2126b89c83fe8b16d4a79a6b303777d8bb264: Status 404 returned error can't find the container with id 721354b0a01b76a735dbaabf45e2126b89c83fe8b16d4a79a6b303777d8bb264 Nov 24 17:36:23 crc kubenswrapper[4808]: I1124 17:36:23.786134 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-cbqfz"] Nov 24 17:36:23 crc kubenswrapper[4808]: W1124 17:36:23.791115 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24eadecc_5782_4e5b_8982_46d3e99fa5b1.slice/crio-3300637ca4869dfa1a9356b010c237074ee969d003aa50d289463a1af5fdd41b WatchSource:0}: Error finding container 3300637ca4869dfa1a9356b010c237074ee969d003aa50d289463a1af5fdd41b: Status 404 returned error can't find the container with id 3300637ca4869dfa1a9356b010c237074ee969d003aa50d289463a1af5fdd41b Nov 24 17:36:24 crc kubenswrapper[4808]: I1124 17:36:24.533365 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-2qsx7" event={"ID":"016a9f20-1203-4241-9210-55b50eeb60a1","Type":"ContainerStarted","Data":"b0291c771a93d83ded9407ab7e78c4f8c465796183ce18106baaefbe9bac4bb0"} Nov 24 17:36:24 crc kubenswrapper[4808]: I1124 17:36:24.535435 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-cbqfz" event={"ID":"24eadecc-5782-4e5b-8982-46d3e99fa5b1","Type":"ContainerStarted","Data":"3300637ca4869dfa1a9356b010c237074ee969d003aa50d289463a1af5fdd41b"} Nov 24 17:36:24 crc kubenswrapper[4808]: I1124 17:36:24.537037 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-bjlpj" event={"ID":"a1ea24f0-1b28-4648-a09e-8c0539792ea1","Type":"ContainerStarted","Data":"721354b0a01b76a735dbaabf45e2126b89c83fe8b16d4a79a6b303777d8bb264"} Nov 24 17:36:32 crc kubenswrapper[4808]: I1124 17:36:32.599772 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-bjlpj" event={"ID":"a1ea24f0-1b28-4648-a09e-8c0539792ea1","Type":"ContainerStarted","Data":"7788705de63de04c07e1a30cbad2591e0505a096980ac914bdc5b5924c3cd7b1"} Nov 24 17:36:32 crc kubenswrapper[4808]: I1124 17:36:32.600680 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-bjlpj" Nov 24 17:36:32 crc kubenswrapper[4808]: I1124 17:36:32.602109 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-2qsx7" event={"ID":"016a9f20-1203-4241-9210-55b50eeb60a1","Type":"ContainerStarted","Data":"00ae5c7c6b1a43f60296fc9488451db5ef46ece0ee9377ff96fad3d99d622aa6"} Nov 24 17:36:32 crc kubenswrapper[4808]: I1124 17:36:32.604649 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-cbqfz" event={"ID":"24eadecc-5782-4e5b-8982-46d3e99fa5b1","Type":"ContainerStarted","Data":"14f5f67f032c9c57fbce4d0a2b82901dcfde91db42422ca21b7bc97b8733d384"} Nov 24 17:36:32 crc kubenswrapper[4808]: I1124 17:36:32.624993 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-bjlpj" podStartSLOduration=1.402926589 podStartE2EDuration="9.624956397s" podCreationTimestamp="2025-11-24 17:36:23 +0000 UTC" firstStartedPulling="2025-11-24 17:36:23.730382006 +0000 UTC m=+576.328049818" lastFinishedPulling="2025-11-24 17:36:31.952411824 +0000 UTC m=+584.550079626" observedRunningTime="2025-11-24 17:36:32.622031494 +0000 UTC m=+585.219699286" watchObservedRunningTime="2025-11-24 17:36:32.624956397 +0000 UTC m=+585.222624219" Nov 24 17:36:32 crc kubenswrapper[4808]: I1124 17:36:32.652208 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-2qsx7" podStartSLOduration=1.452052842 podStartE2EDuration="9.652173594s" podCreationTimestamp="2025-11-24 17:36:23 +0000 UTC" firstStartedPulling="2025-11-24 17:36:23.676884838 +0000 UTC m=+576.274552630" lastFinishedPulling="2025-11-24 17:36:31.87700557 +0000 UTC m=+584.474673382" observedRunningTime="2025-11-24 17:36:32.648125799 +0000 UTC m=+585.245793601" watchObservedRunningTime="2025-11-24 17:36:32.652173594 +0000 UTC m=+585.249841396" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.331162 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-cbqfz" podStartSLOduration=2.248889031 podStartE2EDuration="10.331130739s" podCreationTimestamp="2025-11-24 17:36:23 +0000 UTC" firstStartedPulling="2025-11-24 17:36:23.794762372 +0000 UTC m=+576.392430174" lastFinishedPulling="2025-11-24 17:36:31.87700408 +0000 UTC m=+584.474671882" observedRunningTime="2025-11-24 17:36:32.673095532 +0000 UTC m=+585.270763334" watchObservedRunningTime="2025-11-24 17:36:33.331130739 +0000 UTC m=+585.928798541" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.333363 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5b2z8"] Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.333902 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovn-controller" containerID="cri-o://887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c" gracePeriod=30 Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.333943 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="nbdb" containerID="cri-o://e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481" gracePeriod=30 Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.334062 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="sbdb" containerID="cri-o://cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee" gracePeriod=30 Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.334119 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="kube-rbac-proxy-node" containerID="cri-o://f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24" gracePeriod=30 Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.334140 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d" gracePeriod=30 Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.334192 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="northd" containerID="cri-o://04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672" gracePeriod=30 Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.334218 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovn-acl-logging" containerID="cri-o://920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a" gracePeriod=30 Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.373855 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" containerID="cri-o://f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66" gracePeriod=30 Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.611933 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q5m49_6a192cd6-853d-4643-855d-ac61168b1cda/kube-multus/2.log" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.613538 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q5m49_6a192cd6-853d-4643-855d-ac61168b1cda/kube-multus/1.log" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.613613 4808 generic.go:334] "Generic (PLEG): container finished" podID="6a192cd6-853d-4643-855d-ac61168b1cda" containerID="da95fd35f4059c6d375f4d4a2fd7c5e31b6d03712827a2c6b95cdd1678f55c6b" exitCode=2 Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.613679 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q5m49" event={"ID":"6a192cd6-853d-4643-855d-ac61168b1cda","Type":"ContainerDied","Data":"da95fd35f4059c6d375f4d4a2fd7c5e31b6d03712827a2c6b95cdd1678f55c6b"} Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.613729 4808 scope.go:117] "RemoveContainer" containerID="3ac714762891a8496f5f9fb3d368c22acffd162a1a646871d2be0386c78b0a21" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.614370 4808 scope.go:117] "RemoveContainer" containerID="da95fd35f4059c6d375f4d4a2fd7c5e31b6d03712827a2c6b95cdd1678f55c6b" Nov 24 17:36:33 crc kubenswrapper[4808]: E1124 17:36:33.614727 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-q5m49_openshift-multus(6a192cd6-853d-4643-855d-ac61168b1cda)\"" pod="openshift-multus/multus-q5m49" podUID="6a192cd6-853d-4643-855d-ac61168b1cda" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.618450 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovnkube-controller/3.log" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.622612 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovn-acl-logging/0.log" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.623247 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovn-controller/0.log" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.623645 4808 generic.go:334] "Generic (PLEG): container finished" podID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerID="f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66" exitCode=0 Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.623669 4808 generic.go:334] "Generic (PLEG): container finished" podID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerID="04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672" exitCode=0 Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.623680 4808 generic.go:334] "Generic (PLEG): container finished" podID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerID="793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d" exitCode=0 Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.623689 4808 generic.go:334] "Generic (PLEG): container finished" podID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerID="f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24" exitCode=0 Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.623698 4808 generic.go:334] "Generic (PLEG): container finished" podID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerID="920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a" exitCode=143 Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.623707 4808 generic.go:334] "Generic (PLEG): container finished" podID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerID="887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c" exitCode=143 Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.624415 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerDied","Data":"f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66"} Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.624454 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerDied","Data":"04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672"} Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.624468 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerDied","Data":"793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d"} Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.624481 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerDied","Data":"f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24"} Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.624494 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerDied","Data":"920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a"} Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.624506 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerDied","Data":"887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c"} Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.669883 4808 scope.go:117] "RemoveContainer" containerID="3ce45a29241124b158e3b01317c35744faaad83f2f783f86a9eed729412e2222" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.690305 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovn-acl-logging/0.log" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.690963 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovn-controller/0.log" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.691533 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.747765 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-var-lib-openvswitch\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.747852 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4caad6d6-912a-4441-aa90-38a0e29591f5-ovn-node-metrics-cert\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.747897 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-ovn\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.747947 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-systemd\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.747978 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-ovnkube-script-lib\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748035 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-cni-netd\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748058 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-slash\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748087 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-ovnkube-config\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748112 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-node-log\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748134 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-kubelet\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748153 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-openvswitch\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748180 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-run-netns\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748205 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748244 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-env-overrides\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748270 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-etc-openvswitch\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748302 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-run-ovn-kubernetes\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748322 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-systemd-units\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748346 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-log-socket\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748364 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-cni-bin\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748413 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrc9l\" (UniqueName: \"kubernetes.io/projected/4caad6d6-912a-4441-aa90-38a0e29591f5-kube-api-access-lrc9l\") pod \"4caad6d6-912a-4441-aa90-38a0e29591f5\" (UID: \"4caad6d6-912a-4441-aa90-38a0e29591f5\") " Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748726 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748772 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748814 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748900 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748937 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-slash" (OuterVolumeSpecName: "host-slash") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.748970 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.749041 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.749068 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.749069 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-log-socket" (OuterVolumeSpecName: "log-socket") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.749226 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.749229 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.749255 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.749270 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-node-log" (OuterVolumeSpecName: "node-log") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.749279 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.749303 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.749331 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.750128 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.755828 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4caad6d6-912a-4441-aa90-38a0e29591f5-kube-api-access-lrc9l" (OuterVolumeSpecName: "kube-api-access-lrc9l") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "kube-api-access-lrc9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.756127 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4caad6d6-912a-4441-aa90-38a0e29591f5-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.759759 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-l9gl6"] Nov 24 17:36:33 crc kubenswrapper[4808]: E1124 17:36:33.760054 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760071 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: E1124 17:36:33.760082 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="kube-rbac-proxy-node" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760088 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="kube-rbac-proxy-node" Nov 24 17:36:33 crc kubenswrapper[4808]: E1124 17:36:33.760097 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="nbdb" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760103 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="nbdb" Nov 24 17:36:33 crc kubenswrapper[4808]: E1124 17:36:33.760115 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovn-acl-logging" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760122 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovn-acl-logging" Nov 24 17:36:33 crc kubenswrapper[4808]: E1124 17:36:33.760130 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760136 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 17:36:33 crc kubenswrapper[4808]: E1124 17:36:33.760146 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="northd" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760152 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="northd" Nov 24 17:36:33 crc kubenswrapper[4808]: E1124 17:36:33.760160 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="sbdb" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760167 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="sbdb" Nov 24 17:36:33 crc kubenswrapper[4808]: E1124 17:36:33.760180 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovn-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760186 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovn-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: E1124 17:36:33.760197 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="kubecfg-setup" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760203 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="kubecfg-setup" Nov 24 17:36:33 crc kubenswrapper[4808]: E1124 17:36:33.760211 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760217 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: E1124 17:36:33.760224 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760230 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: E1124 17:36:33.760237 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760242 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760335 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovn-acl-logging" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760342 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760349 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760358 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="sbdb" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760371 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="northd" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760377 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="nbdb" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760385 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760391 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovn-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760397 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="kube-rbac-proxy-node" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760403 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: E1124 17:36:33.760489 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760497 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760590 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.760815 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerName="ovnkube-controller" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.762619 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.772494 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "4caad6d6-912a-4441-aa90-38a0e29591f5" (UID: "4caad6d6-912a-4441-aa90-38a0e29591f5"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850449 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-run-systemd\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850519 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-slash\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850543 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/94cffc5f-1285-4002-8ed7-32deb4bb57ee-env-overrides\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850564 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-cni-bin\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850583 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-run-openvswitch\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850612 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-cni-netd\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850640 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850673 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/94cffc5f-1285-4002-8ed7-32deb4bb57ee-ovnkube-config\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850712 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-node-log\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850747 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptw7c\" (UniqueName: \"kubernetes.io/projected/94cffc5f-1285-4002-8ed7-32deb4bb57ee-kube-api-access-ptw7c\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850774 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-log-socket\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850801 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-kubelet\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850826 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-var-lib-openvswitch\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850861 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-run-netns\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850887 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-run-ovn\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850913 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-etc-openvswitch\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850936 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-run-ovn-kubernetes\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850959 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-systemd-units\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.850985 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/94cffc5f-1285-4002-8ed7-32deb4bb57ee-ovn-node-metrics-cert\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851032 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/94cffc5f-1285-4002-8ed7-32deb4bb57ee-ovnkube-script-lib\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851085 4808 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851101 4808 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851113 4808 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-log-socket\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851126 4808 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851142 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrc9l\" (UniqueName: \"kubernetes.io/projected/4caad6d6-912a-4441-aa90-38a0e29591f5-kube-api-access-lrc9l\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851154 4808 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851167 4808 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4caad6d6-912a-4441-aa90-38a0e29591f5-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851180 4808 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851192 4808 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851205 4808 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851217 4808 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851571 4808 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-slash\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851709 4808 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851749 4808 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851779 4808 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851807 4808 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-node-log\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851836 4808 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851861 4808 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851885 4808 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4caad6d6-912a-4441-aa90-38a0e29591f5-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.851908 4808 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4caad6d6-912a-4441-aa90-38a0e29591f5-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.952782 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-run-netns\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.952831 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-run-ovn\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.952861 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-etc-openvswitch\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.952894 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-run-ovn-kubernetes\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.952917 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-systemd-units\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.952935 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/94cffc5f-1285-4002-8ed7-32deb4bb57ee-ovn-node-metrics-cert\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.952942 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-run-netns\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.952982 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-run-ovn-kubernetes\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.952953 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/94cffc5f-1285-4002-8ed7-32deb4bb57ee-ovnkube-script-lib\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953045 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-systemd-units\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953067 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-run-systemd\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953135 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-slash\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953164 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/94cffc5f-1285-4002-8ed7-32deb4bb57ee-env-overrides\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953195 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-cni-bin\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.952942 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-run-ovn\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953214 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-run-openvswitch\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953239 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-cni-netd\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953245 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-slash\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953265 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953277 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-run-systemd\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953305 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-run-openvswitch\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953331 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/94cffc5f-1285-4002-8ed7-32deb4bb57ee-ovnkube-config\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953345 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-cni-netd\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953392 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953424 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-node-log\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953422 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-cni-bin\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953503 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptw7c\" (UniqueName: \"kubernetes.io/projected/94cffc5f-1285-4002-8ed7-32deb4bb57ee-kube-api-access-ptw7c\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953452 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-node-log\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953536 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-log-socket\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953565 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-kubelet\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953594 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-var-lib-openvswitch\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953674 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-var-lib-openvswitch\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953702 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-log-socket\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953724 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-host-kubelet\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.953754 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/94cffc5f-1285-4002-8ed7-32deb4bb57ee-ovnkube-script-lib\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.954075 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94cffc5f-1285-4002-8ed7-32deb4bb57ee-etc-openvswitch\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.954084 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/94cffc5f-1285-4002-8ed7-32deb4bb57ee-ovnkube-config\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.954301 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/94cffc5f-1285-4002-8ed7-32deb4bb57ee-env-overrides\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.957149 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/94cffc5f-1285-4002-8ed7-32deb4bb57ee-ovn-node-metrics-cert\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:33 crc kubenswrapper[4808]: I1124 17:36:33.973568 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptw7c\" (UniqueName: \"kubernetes.io/projected/94cffc5f-1285-4002-8ed7-32deb4bb57ee-kube-api-access-ptw7c\") pod \"ovnkube-node-l9gl6\" (UID: \"94cffc5f-1285-4002-8ed7-32deb4bb57ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.080553 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:34 crc kubenswrapper[4808]: W1124 17:36:34.101528 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94cffc5f_1285_4002_8ed7_32deb4bb57ee.slice/crio-936936a2bb83470acf245e7ae995ab9ccc32074a7591ae3c9c70ffebc9747e97 WatchSource:0}: Error finding container 936936a2bb83470acf245e7ae995ab9ccc32074a7591ae3c9c70ffebc9747e97: Status 404 returned error can't find the container with id 936936a2bb83470acf245e7ae995ab9ccc32074a7591ae3c9c70ffebc9747e97 Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.637063 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovn-acl-logging/0.log" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.638068 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5b2z8_4caad6d6-912a-4441-aa90-38a0e29591f5/ovn-controller/0.log" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.638518 4808 generic.go:334] "Generic (PLEG): container finished" podID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerID="cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee" exitCode=0 Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.638569 4808 generic.go:334] "Generic (PLEG): container finished" podID="4caad6d6-912a-4441-aa90-38a0e29591f5" containerID="e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481" exitCode=0 Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.638602 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerDied","Data":"cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee"} Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.638684 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.638785 4808 scope.go:117] "RemoveContainer" containerID="f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.638913 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerDied","Data":"e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481"} Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.639381 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5b2z8" event={"ID":"4caad6d6-912a-4441-aa90-38a0e29591f5","Type":"ContainerDied","Data":"e2e107b74a075760d5543d3e2da94b5776dcdda42690920b33b3ca2d8febdac2"} Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.642461 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q5m49_6a192cd6-853d-4643-855d-ac61168b1cda/kube-multus/2.log" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.646287 4808 generic.go:334] "Generic (PLEG): container finished" podID="94cffc5f-1285-4002-8ed7-32deb4bb57ee" containerID="18fffd11e7dfb175f00f83dfc4751ab3599c8c9c379e5e6d5fb485d04c19b24a" exitCode=0 Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.646399 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" event={"ID":"94cffc5f-1285-4002-8ed7-32deb4bb57ee","Type":"ContainerDied","Data":"18fffd11e7dfb175f00f83dfc4751ab3599c8c9c379e5e6d5fb485d04c19b24a"} Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.646466 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" event={"ID":"94cffc5f-1285-4002-8ed7-32deb4bb57ee","Type":"ContainerStarted","Data":"936936a2bb83470acf245e7ae995ab9ccc32074a7591ae3c9c70ffebc9747e97"} Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.658269 4808 scope.go:117] "RemoveContainer" containerID="cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.679828 4808 scope.go:117] "RemoveContainer" containerID="e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.698889 4808 scope.go:117] "RemoveContainer" containerID="04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.728366 4808 scope.go:117] "RemoveContainer" containerID="793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.740424 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5b2z8"] Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.750070 4808 scope.go:117] "RemoveContainer" containerID="f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.753124 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5b2z8"] Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.769149 4808 scope.go:117] "RemoveContainer" containerID="920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.784497 4808 scope.go:117] "RemoveContainer" containerID="887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.799314 4808 scope.go:117] "RemoveContainer" containerID="4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.825179 4808 scope.go:117] "RemoveContainer" containerID="f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66" Nov 24 17:36:34 crc kubenswrapper[4808]: E1124 17:36:34.826228 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66\": container with ID starting with f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66 not found: ID does not exist" containerID="f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.826263 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66"} err="failed to get container status \"f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66\": rpc error: code = NotFound desc = could not find container \"f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66\": container with ID starting with f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66 not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.826290 4808 scope.go:117] "RemoveContainer" containerID="cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee" Nov 24 17:36:34 crc kubenswrapper[4808]: E1124 17:36:34.827937 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\": container with ID starting with cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee not found: ID does not exist" containerID="cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.827964 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee"} err="failed to get container status \"cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\": rpc error: code = NotFound desc = could not find container \"cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\": container with ID starting with cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.827984 4808 scope.go:117] "RemoveContainer" containerID="e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481" Nov 24 17:36:34 crc kubenswrapper[4808]: E1124 17:36:34.830537 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\": container with ID starting with e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481 not found: ID does not exist" containerID="e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.830605 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481"} err="failed to get container status \"e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\": rpc error: code = NotFound desc = could not find container \"e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\": container with ID starting with e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481 not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.830653 4808 scope.go:117] "RemoveContainer" containerID="04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672" Nov 24 17:36:34 crc kubenswrapper[4808]: E1124 17:36:34.831366 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\": container with ID starting with 04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672 not found: ID does not exist" containerID="04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.831419 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672"} err="failed to get container status \"04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\": rpc error: code = NotFound desc = could not find container \"04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\": container with ID starting with 04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672 not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.831438 4808 scope.go:117] "RemoveContainer" containerID="793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d" Nov 24 17:36:34 crc kubenswrapper[4808]: E1124 17:36:34.832094 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\": container with ID starting with 793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d not found: ID does not exist" containerID="793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.832118 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d"} err="failed to get container status \"793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\": rpc error: code = NotFound desc = could not find container \"793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\": container with ID starting with 793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.832133 4808 scope.go:117] "RemoveContainer" containerID="f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24" Nov 24 17:36:34 crc kubenswrapper[4808]: E1124 17:36:34.832539 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\": container with ID starting with f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24 not found: ID does not exist" containerID="f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.832589 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24"} err="failed to get container status \"f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\": rpc error: code = NotFound desc = could not find container \"f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\": container with ID starting with f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24 not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.832604 4808 scope.go:117] "RemoveContainer" containerID="920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a" Nov 24 17:36:34 crc kubenswrapper[4808]: E1124 17:36:34.832904 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\": container with ID starting with 920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a not found: ID does not exist" containerID="920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.832930 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a"} err="failed to get container status \"920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\": rpc error: code = NotFound desc = could not find container \"920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\": container with ID starting with 920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.832946 4808 scope.go:117] "RemoveContainer" containerID="887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c" Nov 24 17:36:34 crc kubenswrapper[4808]: E1124 17:36:34.851873 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\": container with ID starting with 887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c not found: ID does not exist" containerID="887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.851975 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c"} err="failed to get container status \"887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\": rpc error: code = NotFound desc = could not find container \"887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\": container with ID starting with 887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.852007 4808 scope.go:117] "RemoveContainer" containerID="4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09" Nov 24 17:36:34 crc kubenswrapper[4808]: E1124 17:36:34.852601 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\": container with ID starting with 4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09 not found: ID does not exist" containerID="4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.852622 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09"} err="failed to get container status \"4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\": rpc error: code = NotFound desc = could not find container \"4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\": container with ID starting with 4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09 not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.852637 4808 scope.go:117] "RemoveContainer" containerID="f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.853232 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66"} err="failed to get container status \"f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66\": rpc error: code = NotFound desc = could not find container \"f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66\": container with ID starting with f7c37f3460cf8597a7c2b320b4ed2cea848c6915cc54c3229f7eb1ab9b185a66 not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.853313 4808 scope.go:117] "RemoveContainer" containerID="cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.853708 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee"} err="failed to get container status \"cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\": rpc error: code = NotFound desc = could not find container \"cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee\": container with ID starting with cf771233e8bf7348b906aaace4118c904f64f3ead5863466eb76989baf0d92ee not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.853729 4808 scope.go:117] "RemoveContainer" containerID="e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.854036 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481"} err="failed to get container status \"e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\": rpc error: code = NotFound desc = could not find container \"e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481\": container with ID starting with e0aa4cb06f945cde9d8c3c29cfcfdadecf3e0336a7084e7ba784abff84cbe481 not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.854146 4808 scope.go:117] "RemoveContainer" containerID="04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.854476 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672"} err="failed to get container status \"04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\": rpc error: code = NotFound desc = could not find container \"04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672\": container with ID starting with 04831e6c6501e1193cb5ee009206dcc45d75e1a52f1451c9860f7e476c271672 not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.854504 4808 scope.go:117] "RemoveContainer" containerID="793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.854910 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d"} err="failed to get container status \"793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\": rpc error: code = NotFound desc = could not find container \"793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d\": container with ID starting with 793240ba08f89eb1b2c4394f78d02509dc020896a15c37225cd95ae0acf6d73d not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.854955 4808 scope.go:117] "RemoveContainer" containerID="f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.855384 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24"} err="failed to get container status \"f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\": rpc error: code = NotFound desc = could not find container \"f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24\": container with ID starting with f9ab5d6f287e96e4590f54ef81b70f88f8379d3b122ecae9bce24e1c2bdfea24 not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.855405 4808 scope.go:117] "RemoveContainer" containerID="920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.855753 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a"} err="failed to get container status \"920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\": rpc error: code = NotFound desc = could not find container \"920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a\": container with ID starting with 920ff4984d4c693733247dd68f53975f6df5d06c3edc502ba06b574313d05e4a not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.855771 4808 scope.go:117] "RemoveContainer" containerID="887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.856071 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c"} err="failed to get container status \"887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\": rpc error: code = NotFound desc = could not find container \"887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c\": container with ID starting with 887dabf074f95b6d7761a09e8024929f032fb0b97f381dccb2587809fbb0039c not found: ID does not exist" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.856165 4808 scope.go:117] "RemoveContainer" containerID="4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09" Nov 24 17:36:34 crc kubenswrapper[4808]: I1124 17:36:34.856678 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09"} err="failed to get container status \"4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\": rpc error: code = NotFound desc = could not find container \"4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09\": container with ID starting with 4f21b4baab64b05a745dc8c6b90a21ab0cea63ca137ad8fac6801963a773cc09 not found: ID does not exist" Nov 24 17:36:35 crc kubenswrapper[4808]: I1124 17:36:35.656625 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" event={"ID":"94cffc5f-1285-4002-8ed7-32deb4bb57ee","Type":"ContainerStarted","Data":"1e979d477f88de090e868cd953a5261b73587599ba65c44f0a0c051ecd4b28f0"} Nov 24 17:36:35 crc kubenswrapper[4808]: I1124 17:36:35.657190 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" event={"ID":"94cffc5f-1285-4002-8ed7-32deb4bb57ee","Type":"ContainerStarted","Data":"68589be7e41bda0fcd5d81942833934a723055c12c663965c3c7d3cd56f4269c"} Nov 24 17:36:35 crc kubenswrapper[4808]: I1124 17:36:35.657237 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" event={"ID":"94cffc5f-1285-4002-8ed7-32deb4bb57ee","Type":"ContainerStarted","Data":"54320bcda8077cf4b9674b1a3b642ddd467096206e18b770c8bc7d500753ccf1"} Nov 24 17:36:35 crc kubenswrapper[4808]: I1124 17:36:35.657278 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" event={"ID":"94cffc5f-1285-4002-8ed7-32deb4bb57ee","Type":"ContainerStarted","Data":"15d39bd648a28856c53f4957bf9d9e733999c121b3af9fb746e1ef846c63c71d"} Nov 24 17:36:35 crc kubenswrapper[4808]: I1124 17:36:35.657314 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" event={"ID":"94cffc5f-1285-4002-8ed7-32deb4bb57ee","Type":"ContainerStarted","Data":"2199676d3ff3220d5b481449f823383804e37dbafb4b7b651fd25d921513b1b9"} Nov 24 17:36:36 crc kubenswrapper[4808]: I1124 17:36:36.353653 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4caad6d6-912a-4441-aa90-38a0e29591f5" path="/var/lib/kubelet/pods/4caad6d6-912a-4441-aa90-38a0e29591f5/volumes" Nov 24 17:36:36 crc kubenswrapper[4808]: I1124 17:36:36.523191 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:36:36 crc kubenswrapper[4808]: I1124 17:36:36.523257 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:36:36 crc kubenswrapper[4808]: I1124 17:36:36.669882 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" event={"ID":"94cffc5f-1285-4002-8ed7-32deb4bb57ee","Type":"ContainerStarted","Data":"acb96320b0269206eba6e18619be77c0e44da2b2e00e82d6dad4d49b486bab15"} Nov 24 17:36:38 crc kubenswrapper[4808]: I1124 17:36:38.424669 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-bjlpj" Nov 24 17:36:38 crc kubenswrapper[4808]: I1124 17:36:38.695284 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" event={"ID":"94cffc5f-1285-4002-8ed7-32deb4bb57ee","Type":"ContainerStarted","Data":"95d98d7c638593d00a8771ae2c6c8042740785ed5094d79e75572c23ad10c72f"} Nov 24 17:36:40 crc kubenswrapper[4808]: I1124 17:36:40.710925 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" event={"ID":"94cffc5f-1285-4002-8ed7-32deb4bb57ee","Type":"ContainerStarted","Data":"347599a3b441bc58b3a262cada08f791b35007c4ae5d3c951f7127788c7853a8"} Nov 24 17:36:40 crc kubenswrapper[4808]: I1124 17:36:40.746809 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" podStartSLOduration=7.746788545 podStartE2EDuration="7.746788545s" podCreationTimestamp="2025-11-24 17:36:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:36:40.742893864 +0000 UTC m=+593.340561696" watchObservedRunningTime="2025-11-24 17:36:40.746788545 +0000 UTC m=+593.344456347" Nov 24 17:36:41 crc kubenswrapper[4808]: I1124 17:36:41.718892 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:41 crc kubenswrapper[4808]: I1124 17:36:41.718951 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:41 crc kubenswrapper[4808]: I1124 17:36:41.718963 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:41 crc kubenswrapper[4808]: I1124 17:36:41.747813 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:41 crc kubenswrapper[4808]: I1124 17:36:41.749857 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:36:47 crc kubenswrapper[4808]: I1124 17:36:47.346925 4808 scope.go:117] "RemoveContainer" containerID="da95fd35f4059c6d375f4d4a2fd7c5e31b6d03712827a2c6b95cdd1678f55c6b" Nov 24 17:36:47 crc kubenswrapper[4808]: E1124 17:36:47.347849 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-q5m49_openshift-multus(6a192cd6-853d-4643-855d-ac61168b1cda)\"" pod="openshift-multus/multus-q5m49" podUID="6a192cd6-853d-4643-855d-ac61168b1cda" Nov 24 17:36:59 crc kubenswrapper[4808]: I1124 17:36:59.348184 4808 scope.go:117] "RemoveContainer" containerID="da95fd35f4059c6d375f4d4a2fd7c5e31b6d03712827a2c6b95cdd1678f55c6b" Nov 24 17:36:59 crc kubenswrapper[4808]: I1124 17:36:59.854172 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q5m49_6a192cd6-853d-4643-855d-ac61168b1cda/kube-multus/2.log" Nov 24 17:36:59 crc kubenswrapper[4808]: I1124 17:36:59.854643 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q5m49" event={"ID":"6a192cd6-853d-4643-855d-ac61168b1cda","Type":"ContainerStarted","Data":"fefb3e0988fa6bbadbda130e1c474674b524e15bb1f9e6d5bb880e3b1b34d5ad"} Nov 24 17:37:04 crc kubenswrapper[4808]: I1124 17:37:04.106093 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-l9gl6" Nov 24 17:37:06 crc kubenswrapper[4808]: I1124 17:37:06.523343 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:37:06 crc kubenswrapper[4808]: I1124 17:37:06.523946 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:37:22 crc kubenswrapper[4808]: I1124 17:37:22.430614 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp"] Nov 24 17:37:22 crc kubenswrapper[4808]: I1124 17:37:22.433088 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" Nov 24 17:37:22 crc kubenswrapper[4808]: I1124 17:37:22.437334 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 17:37:22 crc kubenswrapper[4808]: I1124 17:37:22.447520 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp"] Nov 24 17:37:22 crc kubenswrapper[4808]: I1124 17:37:22.453769 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/432d4d4f-23e9-43f7-bf64-af0caa59dbea-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp\" (UID: \"432d4d4f-23e9-43f7-bf64-af0caa59dbea\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" Nov 24 17:37:22 crc kubenswrapper[4808]: I1124 17:37:22.453935 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb5mp\" (UniqueName: \"kubernetes.io/projected/432d4d4f-23e9-43f7-bf64-af0caa59dbea-kube-api-access-pb5mp\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp\" (UID: \"432d4d4f-23e9-43f7-bf64-af0caa59dbea\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" Nov 24 17:37:22 crc kubenswrapper[4808]: I1124 17:37:22.454037 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/432d4d4f-23e9-43f7-bf64-af0caa59dbea-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp\" (UID: \"432d4d4f-23e9-43f7-bf64-af0caa59dbea\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" Nov 24 17:37:22 crc kubenswrapper[4808]: I1124 17:37:22.555033 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/432d4d4f-23e9-43f7-bf64-af0caa59dbea-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp\" (UID: \"432d4d4f-23e9-43f7-bf64-af0caa59dbea\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" Nov 24 17:37:22 crc kubenswrapper[4808]: I1124 17:37:22.555116 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb5mp\" (UniqueName: \"kubernetes.io/projected/432d4d4f-23e9-43f7-bf64-af0caa59dbea-kube-api-access-pb5mp\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp\" (UID: \"432d4d4f-23e9-43f7-bf64-af0caa59dbea\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" Nov 24 17:37:22 crc kubenswrapper[4808]: I1124 17:37:22.555154 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/432d4d4f-23e9-43f7-bf64-af0caa59dbea-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp\" (UID: \"432d4d4f-23e9-43f7-bf64-af0caa59dbea\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" Nov 24 17:37:22 crc kubenswrapper[4808]: I1124 17:37:22.555614 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/432d4d4f-23e9-43f7-bf64-af0caa59dbea-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp\" (UID: \"432d4d4f-23e9-43f7-bf64-af0caa59dbea\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" Nov 24 17:37:22 crc kubenswrapper[4808]: I1124 17:37:22.555696 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/432d4d4f-23e9-43f7-bf64-af0caa59dbea-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp\" (UID: \"432d4d4f-23e9-43f7-bf64-af0caa59dbea\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" Nov 24 17:37:22 crc kubenswrapper[4808]: I1124 17:37:22.578177 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb5mp\" (UniqueName: \"kubernetes.io/projected/432d4d4f-23e9-43f7-bf64-af0caa59dbea-kube-api-access-pb5mp\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp\" (UID: \"432d4d4f-23e9-43f7-bf64-af0caa59dbea\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" Nov 24 17:37:22 crc kubenswrapper[4808]: I1124 17:37:22.757007 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" Nov 24 17:37:22 crc kubenswrapper[4808]: I1124 17:37:22.970526 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp"] Nov 24 17:37:23 crc kubenswrapper[4808]: I1124 17:37:23.006713 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" event={"ID":"432d4d4f-23e9-43f7-bf64-af0caa59dbea","Type":"ContainerStarted","Data":"a9c4c2f4958004e1e4cdeb9df2a2e0728522a5feef0e568103d97945f6091350"} Nov 24 17:37:24 crc kubenswrapper[4808]: I1124 17:37:24.014267 4808 generic.go:334] "Generic (PLEG): container finished" podID="432d4d4f-23e9-43f7-bf64-af0caa59dbea" containerID="a72b862c9ffd31601ea7d6af46a6ceb093e543ad130047eecb014c3b07fe17c8" exitCode=0 Nov 24 17:37:24 crc kubenswrapper[4808]: I1124 17:37:24.014318 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" event={"ID":"432d4d4f-23e9-43f7-bf64-af0caa59dbea","Type":"ContainerDied","Data":"a72b862c9ffd31601ea7d6af46a6ceb093e543ad130047eecb014c3b07fe17c8"} Nov 24 17:37:26 crc kubenswrapper[4808]: I1124 17:37:26.029295 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" event={"ID":"432d4d4f-23e9-43f7-bf64-af0caa59dbea","Type":"ContainerStarted","Data":"654f4d309fa878dd42d849fa05f274de84a6280cba21c75b3f2b2ef5ed7b91d8"} Nov 24 17:37:27 crc kubenswrapper[4808]: I1124 17:37:27.037356 4808 generic.go:334] "Generic (PLEG): container finished" podID="432d4d4f-23e9-43f7-bf64-af0caa59dbea" containerID="654f4d309fa878dd42d849fa05f274de84a6280cba21c75b3f2b2ef5ed7b91d8" exitCode=0 Nov 24 17:37:27 crc kubenswrapper[4808]: I1124 17:37:27.037468 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" event={"ID":"432d4d4f-23e9-43f7-bf64-af0caa59dbea","Type":"ContainerDied","Data":"654f4d309fa878dd42d849fa05f274de84a6280cba21c75b3f2b2ef5ed7b91d8"} Nov 24 17:37:28 crc kubenswrapper[4808]: I1124 17:37:28.049475 4808 generic.go:334] "Generic (PLEG): container finished" podID="432d4d4f-23e9-43f7-bf64-af0caa59dbea" containerID="9ef679b7d3f26879395f97c55ce434ef94b85cf0ef9e817c82724f6dc102b5b0" exitCode=0 Nov 24 17:37:28 crc kubenswrapper[4808]: I1124 17:37:28.049528 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" event={"ID":"432d4d4f-23e9-43f7-bf64-af0caa59dbea","Type":"ContainerDied","Data":"9ef679b7d3f26879395f97c55ce434ef94b85cf0ef9e817c82724f6dc102b5b0"} Nov 24 17:37:29 crc kubenswrapper[4808]: I1124 17:37:29.293929 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" Nov 24 17:37:29 crc kubenswrapper[4808]: I1124 17:37:29.354740 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/432d4d4f-23e9-43f7-bf64-af0caa59dbea-bundle\") pod \"432d4d4f-23e9-43f7-bf64-af0caa59dbea\" (UID: \"432d4d4f-23e9-43f7-bf64-af0caa59dbea\") " Nov 24 17:37:29 crc kubenswrapper[4808]: I1124 17:37:29.354855 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/432d4d4f-23e9-43f7-bf64-af0caa59dbea-util\") pod \"432d4d4f-23e9-43f7-bf64-af0caa59dbea\" (UID: \"432d4d4f-23e9-43f7-bf64-af0caa59dbea\") " Nov 24 17:37:29 crc kubenswrapper[4808]: I1124 17:37:29.354918 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb5mp\" (UniqueName: \"kubernetes.io/projected/432d4d4f-23e9-43f7-bf64-af0caa59dbea-kube-api-access-pb5mp\") pod \"432d4d4f-23e9-43f7-bf64-af0caa59dbea\" (UID: \"432d4d4f-23e9-43f7-bf64-af0caa59dbea\") " Nov 24 17:37:29 crc kubenswrapper[4808]: I1124 17:37:29.355808 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/432d4d4f-23e9-43f7-bf64-af0caa59dbea-bundle" (OuterVolumeSpecName: "bundle") pod "432d4d4f-23e9-43f7-bf64-af0caa59dbea" (UID: "432d4d4f-23e9-43f7-bf64-af0caa59dbea"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:37:29 crc kubenswrapper[4808]: I1124 17:37:29.365791 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/432d4d4f-23e9-43f7-bf64-af0caa59dbea-kube-api-access-pb5mp" (OuterVolumeSpecName: "kube-api-access-pb5mp") pod "432d4d4f-23e9-43f7-bf64-af0caa59dbea" (UID: "432d4d4f-23e9-43f7-bf64-af0caa59dbea"). InnerVolumeSpecName "kube-api-access-pb5mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:37:29 crc kubenswrapper[4808]: I1124 17:37:29.369329 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/432d4d4f-23e9-43f7-bf64-af0caa59dbea-util" (OuterVolumeSpecName: "util") pod "432d4d4f-23e9-43f7-bf64-af0caa59dbea" (UID: "432d4d4f-23e9-43f7-bf64-af0caa59dbea"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:37:29 crc kubenswrapper[4808]: I1124 17:37:29.457243 4808 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/432d4d4f-23e9-43f7-bf64-af0caa59dbea-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:37:29 crc kubenswrapper[4808]: I1124 17:37:29.457316 4808 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/432d4d4f-23e9-43f7-bf64-af0caa59dbea-util\") on node \"crc\" DevicePath \"\"" Nov 24 17:37:29 crc kubenswrapper[4808]: I1124 17:37:29.457333 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb5mp\" (UniqueName: \"kubernetes.io/projected/432d4d4f-23e9-43f7-bf64-af0caa59dbea-kube-api-access-pb5mp\") on node \"crc\" DevicePath \"\"" Nov 24 17:37:30 crc kubenswrapper[4808]: I1124 17:37:30.065134 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" event={"ID":"432d4d4f-23e9-43f7-bf64-af0caa59dbea","Type":"ContainerDied","Data":"a9c4c2f4958004e1e4cdeb9df2a2e0728522a5feef0e568103d97945f6091350"} Nov 24 17:37:30 crc kubenswrapper[4808]: I1124 17:37:30.065533 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9c4c2f4958004e1e4cdeb9df2a2e0728522a5feef0e568103d97945f6091350" Nov 24 17:37:30 crc kubenswrapper[4808]: I1124 17:37:30.065692 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp" Nov 24 17:37:33 crc kubenswrapper[4808]: I1124 17:37:33.972439 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-6dp5q"] Nov 24 17:37:33 crc kubenswrapper[4808]: E1124 17:37:33.974127 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="432d4d4f-23e9-43f7-bf64-af0caa59dbea" containerName="extract" Nov 24 17:37:33 crc kubenswrapper[4808]: I1124 17:37:33.974220 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="432d4d4f-23e9-43f7-bf64-af0caa59dbea" containerName="extract" Nov 24 17:37:33 crc kubenswrapper[4808]: E1124 17:37:33.974323 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="432d4d4f-23e9-43f7-bf64-af0caa59dbea" containerName="util" Nov 24 17:37:33 crc kubenswrapper[4808]: I1124 17:37:33.974394 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="432d4d4f-23e9-43f7-bf64-af0caa59dbea" containerName="util" Nov 24 17:37:33 crc kubenswrapper[4808]: E1124 17:37:33.974466 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="432d4d4f-23e9-43f7-bf64-af0caa59dbea" containerName="pull" Nov 24 17:37:33 crc kubenswrapper[4808]: I1124 17:37:33.974531 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="432d4d4f-23e9-43f7-bf64-af0caa59dbea" containerName="pull" Nov 24 17:37:33 crc kubenswrapper[4808]: I1124 17:37:33.974717 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="432d4d4f-23e9-43f7-bf64-af0caa59dbea" containerName="extract" Nov 24 17:37:33 crc kubenswrapper[4808]: I1124 17:37:33.975362 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-6dp5q" Nov 24 17:37:33 crc kubenswrapper[4808]: I1124 17:37:33.977751 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-vbrlc" Nov 24 17:37:33 crc kubenswrapper[4808]: I1124 17:37:33.978984 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 24 17:37:33 crc kubenswrapper[4808]: I1124 17:37:33.983149 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 24 17:37:33 crc kubenswrapper[4808]: I1124 17:37:33.987797 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-6dp5q"] Nov 24 17:37:34 crc kubenswrapper[4808]: I1124 17:37:34.021433 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdw6w\" (UniqueName: \"kubernetes.io/projected/48267952-ed56-4a20-98c9-e22f04e52c63-kube-api-access-gdw6w\") pod \"nmstate-operator-557fdffb88-6dp5q\" (UID: \"48267952-ed56-4a20-98c9-e22f04e52c63\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-6dp5q" Nov 24 17:37:34 crc kubenswrapper[4808]: I1124 17:37:34.123141 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdw6w\" (UniqueName: \"kubernetes.io/projected/48267952-ed56-4a20-98c9-e22f04e52c63-kube-api-access-gdw6w\") pod \"nmstate-operator-557fdffb88-6dp5q\" (UID: \"48267952-ed56-4a20-98c9-e22f04e52c63\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-6dp5q" Nov 24 17:37:34 crc kubenswrapper[4808]: I1124 17:37:34.148481 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdw6w\" (UniqueName: \"kubernetes.io/projected/48267952-ed56-4a20-98c9-e22f04e52c63-kube-api-access-gdw6w\") pod \"nmstate-operator-557fdffb88-6dp5q\" (UID: \"48267952-ed56-4a20-98c9-e22f04e52c63\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-6dp5q" Nov 24 17:37:34 crc kubenswrapper[4808]: I1124 17:37:34.297339 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-6dp5q" Nov 24 17:37:34 crc kubenswrapper[4808]: I1124 17:37:34.529352 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-6dp5q"] Nov 24 17:37:35 crc kubenswrapper[4808]: I1124 17:37:35.102258 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-6dp5q" event={"ID":"48267952-ed56-4a20-98c9-e22f04e52c63","Type":"ContainerStarted","Data":"d4bc01056352b1b6f0bd64c35a05ab87f9936638f044c38bdbd8285133be6b6f"} Nov 24 17:37:36 crc kubenswrapper[4808]: I1124 17:37:36.523578 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:37:36 crc kubenswrapper[4808]: I1124 17:37:36.524234 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:37:36 crc kubenswrapper[4808]: I1124 17:37:36.524353 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:37:36 crc kubenswrapper[4808]: I1124 17:37:36.525934 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8b5c11a71abbce839aaa56f6f8510529f68981025ad2c3fda226177d751485f2"} pod="openshift-machine-config-operator/machine-config-daemon-ht87w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:37:36 crc kubenswrapper[4808]: I1124 17:37:36.526136 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" containerID="cri-o://8b5c11a71abbce839aaa56f6f8510529f68981025ad2c3fda226177d751485f2" gracePeriod=600 Nov 24 17:37:37 crc kubenswrapper[4808]: I1124 17:37:37.120625 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-6dp5q" event={"ID":"48267952-ed56-4a20-98c9-e22f04e52c63","Type":"ContainerStarted","Data":"7f6ce318bf15c92783ecac33ac524f31ae22f4cda251d24b66d92dc6d72ad951"} Nov 24 17:37:37 crc kubenswrapper[4808]: I1124 17:37:37.124002 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerID="8b5c11a71abbce839aaa56f6f8510529f68981025ad2c3fda226177d751485f2" exitCode=0 Nov 24 17:37:37 crc kubenswrapper[4808]: I1124 17:37:37.124055 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerDied","Data":"8b5c11a71abbce839aaa56f6f8510529f68981025ad2c3fda226177d751485f2"} Nov 24 17:37:37 crc kubenswrapper[4808]: I1124 17:37:37.124136 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"36f63fea078d49c365c55146d216a74b13586224c5d10cc6c372c7c5a5288461"} Nov 24 17:37:37 crc kubenswrapper[4808]: I1124 17:37:37.124179 4808 scope.go:117] "RemoveContainer" containerID="4b9523f5392391855bf06e84ed62910c36d9c20fb3584ca98f4425fec2b9f41c" Nov 24 17:37:37 crc kubenswrapper[4808]: I1124 17:37:37.142460 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-6dp5q" podStartSLOduration=2.090293791 podStartE2EDuration="4.142423366s" podCreationTimestamp="2025-11-24 17:37:33 +0000 UTC" firstStartedPulling="2025-11-24 17:37:34.53538925 +0000 UTC m=+647.133057052" lastFinishedPulling="2025-11-24 17:37:36.587518825 +0000 UTC m=+649.185186627" observedRunningTime="2025-11-24 17:37:37.13964432 +0000 UTC m=+649.737312122" watchObservedRunningTime="2025-11-24 17:37:37.142423366 +0000 UTC m=+649.740091168" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.717192 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-mf5qh"] Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.719100 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-mf5qh" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.721398 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-j9czb" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.733808 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm"] Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.735465 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.738814 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.742549 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-mf5qh"] Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.760550 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm"] Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.774792 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-kz6lk"] Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.775550 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.854235 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b1fd2203-60a6-4172-bb4e-124fc185ade8-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-rzhnm\" (UID: \"b1fd2203-60a6-4172-bb4e-124fc185ade8\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.854706 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p65b4\" (UniqueName: \"kubernetes.io/projected/125495c2-7b26-44c8-913a-0b98c6c81c98-kube-api-access-p65b4\") pod \"nmstate-metrics-5dcf9c57c5-mf5qh\" (UID: \"125495c2-7b26-44c8-913a-0b98c6c81c98\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-mf5qh" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.854875 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt2hj\" (UniqueName: \"kubernetes.io/projected/9cc7b294-4f80-47be-88f4-e3b0c010ca1d-kube-api-access-mt2hj\") pod \"nmstate-handler-kz6lk\" (UID: \"9cc7b294-4f80-47be-88f4-e3b0c010ca1d\") " pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.855050 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mq9x\" (UniqueName: \"kubernetes.io/projected/b1fd2203-60a6-4172-bb4e-124fc185ade8-kube-api-access-8mq9x\") pod \"nmstate-webhook-6b89b748d8-rzhnm\" (UID: \"b1fd2203-60a6-4172-bb4e-124fc185ade8\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.855175 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/9cc7b294-4f80-47be-88f4-e3b0c010ca1d-dbus-socket\") pod \"nmstate-handler-kz6lk\" (UID: \"9cc7b294-4f80-47be-88f4-e3b0c010ca1d\") " pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.855319 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/9cc7b294-4f80-47be-88f4-e3b0c010ca1d-nmstate-lock\") pod \"nmstate-handler-kz6lk\" (UID: \"9cc7b294-4f80-47be-88f4-e3b0c010ca1d\") " pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.855494 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/9cc7b294-4f80-47be-88f4-e3b0c010ca1d-ovs-socket\") pod \"nmstate-handler-kz6lk\" (UID: \"9cc7b294-4f80-47be-88f4-e3b0c010ca1d\") " pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.957446 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/9cc7b294-4f80-47be-88f4-e3b0c010ca1d-nmstate-lock\") pod \"nmstate-handler-kz6lk\" (UID: \"9cc7b294-4f80-47be-88f4-e3b0c010ca1d\") " pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.957800 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/9cc7b294-4f80-47be-88f4-e3b0c010ca1d-ovs-socket\") pod \"nmstate-handler-kz6lk\" (UID: \"9cc7b294-4f80-47be-88f4-e3b0c010ca1d\") " pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.957944 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b1fd2203-60a6-4172-bb4e-124fc185ade8-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-rzhnm\" (UID: \"b1fd2203-60a6-4172-bb4e-124fc185ade8\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.958069 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p65b4\" (UniqueName: \"kubernetes.io/projected/125495c2-7b26-44c8-913a-0b98c6c81c98-kube-api-access-p65b4\") pod \"nmstate-metrics-5dcf9c57c5-mf5qh\" (UID: \"125495c2-7b26-44c8-913a-0b98c6c81c98\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-mf5qh" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.958188 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt2hj\" (UniqueName: \"kubernetes.io/projected/9cc7b294-4f80-47be-88f4-e3b0c010ca1d-kube-api-access-mt2hj\") pod \"nmstate-handler-kz6lk\" (UID: \"9cc7b294-4f80-47be-88f4-e3b0c010ca1d\") " pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.958287 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mq9x\" (UniqueName: \"kubernetes.io/projected/b1fd2203-60a6-4172-bb4e-124fc185ade8-kube-api-access-8mq9x\") pod \"nmstate-webhook-6b89b748d8-rzhnm\" (UID: \"b1fd2203-60a6-4172-bb4e-124fc185ade8\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.958369 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/9cc7b294-4f80-47be-88f4-e3b0c010ca1d-dbus-socket\") pod \"nmstate-handler-kz6lk\" (UID: \"9cc7b294-4f80-47be-88f4-e3b0c010ca1d\") " pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.957654 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/9cc7b294-4f80-47be-88f4-e3b0c010ca1d-nmstate-lock\") pod \"nmstate-handler-kz6lk\" (UID: \"9cc7b294-4f80-47be-88f4-e3b0c010ca1d\") " pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.957863 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/9cc7b294-4f80-47be-88f4-e3b0c010ca1d-ovs-socket\") pod \"nmstate-handler-kz6lk\" (UID: \"9cc7b294-4f80-47be-88f4-e3b0c010ca1d\") " pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:42 crc kubenswrapper[4808]: E1124 17:37:42.958102 4808 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Nov 24 17:37:42 crc kubenswrapper[4808]: E1124 17:37:42.958704 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1fd2203-60a6-4172-bb4e-124fc185ade8-tls-key-pair podName:b1fd2203-60a6-4172-bb4e-124fc185ade8 nodeName:}" failed. No retries permitted until 2025-11-24 17:37:43.458676926 +0000 UTC m=+656.056344728 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/b1fd2203-60a6-4172-bb4e-124fc185ade8-tls-key-pair") pod "nmstate-webhook-6b89b748d8-rzhnm" (UID: "b1fd2203-60a6-4172-bb4e-124fc185ade8") : secret "openshift-nmstate-webhook" not found Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.958962 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/9cc7b294-4f80-47be-88f4-e3b0c010ca1d-dbus-socket\") pod \"nmstate-handler-kz6lk\" (UID: \"9cc7b294-4f80-47be-88f4-e3b0c010ca1d\") " pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:42 crc kubenswrapper[4808]: I1124 17:37:42.981226 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c"] Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:42.982433 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:42.984263 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p65b4\" (UniqueName: \"kubernetes.io/projected/125495c2-7b26-44c8-913a-0b98c6c81c98-kube-api-access-p65b4\") pod \"nmstate-metrics-5dcf9c57c5-mf5qh\" (UID: \"125495c2-7b26-44c8-913a-0b98c6c81c98\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-mf5qh" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:42.985736 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt2hj\" (UniqueName: \"kubernetes.io/projected/9cc7b294-4f80-47be-88f4-e3b0c010ca1d-kube-api-access-mt2hj\") pod \"nmstate-handler-kz6lk\" (UID: \"9cc7b294-4f80-47be-88f4-e3b0c010ca1d\") " pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:42.998432 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c"] Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.008998 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mq9x\" (UniqueName: \"kubernetes.io/projected/b1fd2203-60a6-4172-bb4e-124fc185ade8-kube-api-access-8mq9x\") pod \"nmstate-webhook-6b89b748d8-rzhnm\" (UID: \"b1fd2203-60a6-4172-bb4e-124fc185ade8\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.024543 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.024794 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-z8hvt" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.024982 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.042061 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-mf5qh" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.060141 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z7bz\" (UniqueName: \"kubernetes.io/projected/b9930406-6bd1-4674-a697-e915c209fc08-kube-api-access-7z7bz\") pod \"nmstate-console-plugin-5874bd7bc5-4x67c\" (UID: \"b9930406-6bd1-4674-a697-e915c209fc08\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.060251 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b9930406-6bd1-4674-a697-e915c209fc08-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-4x67c\" (UID: \"b9930406-6bd1-4674-a697-e915c209fc08\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.060274 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b9930406-6bd1-4674-a697-e915c209fc08-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-4x67c\" (UID: \"b9930406-6bd1-4674-a697-e915c209fc08\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.100752 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.162810 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b9930406-6bd1-4674-a697-e915c209fc08-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-4x67c\" (UID: \"b9930406-6bd1-4674-a697-e915c209fc08\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.163540 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b9930406-6bd1-4674-a697-e915c209fc08-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-4x67c\" (UID: \"b9930406-6bd1-4674-a697-e915c209fc08\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c" Nov 24 17:37:43 crc kubenswrapper[4808]: E1124 17:37:43.163633 4808 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.163670 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z7bz\" (UniqueName: \"kubernetes.io/projected/b9930406-6bd1-4674-a697-e915c209fc08-kube-api-access-7z7bz\") pod \"nmstate-console-plugin-5874bd7bc5-4x67c\" (UID: \"b9930406-6bd1-4674-a697-e915c209fc08\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c" Nov 24 17:37:43 crc kubenswrapper[4808]: E1124 17:37:43.163768 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b9930406-6bd1-4674-a697-e915c209fc08-plugin-serving-cert podName:b9930406-6bd1-4674-a697-e915c209fc08 nodeName:}" failed. No retries permitted until 2025-11-24 17:37:43.663733844 +0000 UTC m=+656.261401646 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/b9930406-6bd1-4674-a697-e915c209fc08-plugin-serving-cert") pod "nmstate-console-plugin-5874bd7bc5-4x67c" (UID: "b9930406-6bd1-4674-a697-e915c209fc08") : secret "plugin-serving-cert" not found Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.166048 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b9930406-6bd1-4674-a697-e915c209fc08-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-4x67c\" (UID: \"b9930406-6bd1-4674-a697-e915c209fc08\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.172994 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-kz6lk" event={"ID":"9cc7b294-4f80-47be-88f4-e3b0c010ca1d","Type":"ContainerStarted","Data":"4fd058d7b942942f2f0718a64616bdf8c56184e91703bff72d33a53ac96ab1ee"} Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.210181 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z7bz\" (UniqueName: \"kubernetes.io/projected/b9930406-6bd1-4674-a697-e915c209fc08-kube-api-access-7z7bz\") pod \"nmstate-console-plugin-5874bd7bc5-4x67c\" (UID: \"b9930406-6bd1-4674-a697-e915c209fc08\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.218224 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-76fd5f8594-gndtm"] Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.219171 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.232362 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-76fd5f8594-gndtm"] Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.368781 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ded92b5f-e571-422f-9f37-6e501be16b4d-service-ca\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.368852 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ded92b5f-e571-422f-9f37-6e501be16b4d-console-oauth-config\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.368881 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ded92b5f-e571-422f-9f37-6e501be16b4d-oauth-serving-cert\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.368947 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ded92b5f-e571-422f-9f37-6e501be16b4d-console-config\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.369077 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fn96\" (UniqueName: \"kubernetes.io/projected/ded92b5f-e571-422f-9f37-6e501be16b4d-kube-api-access-5fn96\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.369213 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ded92b5f-e571-422f-9f37-6e501be16b4d-trusted-ca-bundle\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.369469 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ded92b5f-e571-422f-9f37-6e501be16b4d-console-serving-cert\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.470942 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b1fd2203-60a6-4172-bb4e-124fc185ade8-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-rzhnm\" (UID: \"b1fd2203-60a6-4172-bb4e-124fc185ade8\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.471031 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ded92b5f-e571-422f-9f37-6e501be16b4d-service-ca\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.471072 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ded92b5f-e571-422f-9f37-6e501be16b4d-console-oauth-config\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.471101 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ded92b5f-e571-422f-9f37-6e501be16b4d-oauth-serving-cert\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.471157 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ded92b5f-e571-422f-9f37-6e501be16b4d-console-config\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.471178 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fn96\" (UniqueName: \"kubernetes.io/projected/ded92b5f-e571-422f-9f37-6e501be16b4d-kube-api-access-5fn96\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.471206 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ded92b5f-e571-422f-9f37-6e501be16b4d-trusted-ca-bundle\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.471277 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ded92b5f-e571-422f-9f37-6e501be16b4d-console-serving-cert\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.473259 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ded92b5f-e571-422f-9f37-6e501be16b4d-service-ca\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.473433 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ded92b5f-e571-422f-9f37-6e501be16b4d-console-config\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.473624 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ded92b5f-e571-422f-9f37-6e501be16b4d-trusted-ca-bundle\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.474304 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ded92b5f-e571-422f-9f37-6e501be16b4d-oauth-serving-cert\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.475541 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ded92b5f-e571-422f-9f37-6e501be16b4d-console-serving-cert\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.475550 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b1fd2203-60a6-4172-bb4e-124fc185ade8-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-rzhnm\" (UID: \"b1fd2203-60a6-4172-bb4e-124fc185ade8\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.481714 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ded92b5f-e571-422f-9f37-6e501be16b4d-console-oauth-config\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.493645 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fn96\" (UniqueName: \"kubernetes.io/projected/ded92b5f-e571-422f-9f37-6e501be16b4d-kube-api-access-5fn96\") pod \"console-76fd5f8594-gndtm\" (UID: \"ded92b5f-e571-422f-9f37-6e501be16b4d\") " pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.548066 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.611441 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-mf5qh"] Nov 24 17:37:43 crc kubenswrapper[4808]: W1124 17:37:43.616570 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod125495c2_7b26_44c8_913a_0b98c6c81c98.slice/crio-645730c9cfa48919ec63afc8cddbc82e401c5db33308f1c5c4a6e3837c65bcf8 WatchSource:0}: Error finding container 645730c9cfa48919ec63afc8cddbc82e401c5db33308f1c5c4a6e3837c65bcf8: Status 404 returned error can't find the container with id 645730c9cfa48919ec63afc8cddbc82e401c5db33308f1c5c4a6e3837c65bcf8 Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.662601 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.673933 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b9930406-6bd1-4674-a697-e915c209fc08-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-4x67c\" (UID: \"b9930406-6bd1-4674-a697-e915c209fc08\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.678975 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b9930406-6bd1-4674-a697-e915c209fc08-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-4x67c\" (UID: \"b9930406-6bd1-4674-a697-e915c209fc08\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.687583 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c" Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.767258 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-76fd5f8594-gndtm"] Nov 24 17:37:43 crc kubenswrapper[4808]: W1124 17:37:43.778286 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podded92b5f_e571_422f_9f37_6e501be16b4d.slice/crio-d828a746c2b57b0083c7bbba0521eac9bbe79fe3380d1dcb79fc582569f5bf40 WatchSource:0}: Error finding container d828a746c2b57b0083c7bbba0521eac9bbe79fe3380d1dcb79fc582569f5bf40: Status 404 returned error can't find the container with id d828a746c2b57b0083c7bbba0521eac9bbe79fe3380d1dcb79fc582569f5bf40 Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.954121 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm"] Nov 24 17:37:43 crc kubenswrapper[4808]: I1124 17:37:43.996237 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c"] Nov 24 17:37:44 crc kubenswrapper[4808]: I1124 17:37:44.179941 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c" event={"ID":"b9930406-6bd1-4674-a697-e915c209fc08","Type":"ContainerStarted","Data":"795b12e55c5e23855b0a8f84694e7bcb09fbfc8eaff2585e7595691f2c60e38f"} Nov 24 17:37:44 crc kubenswrapper[4808]: I1124 17:37:44.181986 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-76fd5f8594-gndtm" event={"ID":"ded92b5f-e571-422f-9f37-6e501be16b4d","Type":"ContainerStarted","Data":"f36d83ef331aeaf86afea61a8a3109547e2cdefc95ceb9933c76f6444c205b2e"} Nov 24 17:37:44 crc kubenswrapper[4808]: I1124 17:37:44.182052 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-76fd5f8594-gndtm" event={"ID":"ded92b5f-e571-422f-9f37-6e501be16b4d","Type":"ContainerStarted","Data":"d828a746c2b57b0083c7bbba0521eac9bbe79fe3380d1dcb79fc582569f5bf40"} Nov 24 17:37:44 crc kubenswrapper[4808]: I1124 17:37:44.184045 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-mf5qh" event={"ID":"125495c2-7b26-44c8-913a-0b98c6c81c98","Type":"ContainerStarted","Data":"645730c9cfa48919ec63afc8cddbc82e401c5db33308f1c5c4a6e3837c65bcf8"} Nov 24 17:37:44 crc kubenswrapper[4808]: I1124 17:37:44.184866 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm" event={"ID":"b1fd2203-60a6-4172-bb4e-124fc185ade8","Type":"ContainerStarted","Data":"826554ed10e10e82c70ac2f2fd42ee455abb9580c00259a09d565893e64feaf0"} Nov 24 17:37:44 crc kubenswrapper[4808]: I1124 17:37:44.201444 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-76fd5f8594-gndtm" podStartSLOduration=1.20140825 podStartE2EDuration="1.20140825s" podCreationTimestamp="2025-11-24 17:37:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:37:44.200493225 +0000 UTC m=+656.798161027" watchObservedRunningTime="2025-11-24 17:37:44.20140825 +0000 UTC m=+656.799076042" Nov 24 17:37:46 crc kubenswrapper[4808]: I1124 17:37:46.205160 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-kz6lk" event={"ID":"9cc7b294-4f80-47be-88f4-e3b0c010ca1d","Type":"ContainerStarted","Data":"26f9edbacc43432bf129a20c78b0a0f86dbb627fcfc553f855a2557050914de0"} Nov 24 17:37:46 crc kubenswrapper[4808]: I1124 17:37:46.205710 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:46 crc kubenswrapper[4808]: I1124 17:37:46.207691 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm" event={"ID":"b1fd2203-60a6-4172-bb4e-124fc185ade8","Type":"ContainerStarted","Data":"9dd730e23a38e5110b5d7e7e5a395a9a5f5f0a1b4e21f221bd0b073e9fdbbb31"} Nov 24 17:37:46 crc kubenswrapper[4808]: I1124 17:37:46.208665 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm" Nov 24 17:37:46 crc kubenswrapper[4808]: I1124 17:37:46.211068 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-mf5qh" event={"ID":"125495c2-7b26-44c8-913a-0b98c6c81c98","Type":"ContainerStarted","Data":"aeeeb5ae63b554321f89c145a884e3382d8b19e2214e8122136bd9d6a8fd2153"} Nov 24 17:37:46 crc kubenswrapper[4808]: I1124 17:37:46.223795 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-kz6lk" podStartSLOduration=1.641708937 podStartE2EDuration="4.22377231s" podCreationTimestamp="2025-11-24 17:37:42 +0000 UTC" firstStartedPulling="2025-11-24 17:37:43.142099901 +0000 UTC m=+655.739767713" lastFinishedPulling="2025-11-24 17:37:45.724163274 +0000 UTC m=+658.321831086" observedRunningTime="2025-11-24 17:37:46.220544042 +0000 UTC m=+658.818211854" watchObservedRunningTime="2025-11-24 17:37:46.22377231 +0000 UTC m=+658.821440112" Nov 24 17:37:46 crc kubenswrapper[4808]: I1124 17:37:46.251386 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm" podStartSLOduration=2.471948421 podStartE2EDuration="4.251348946s" podCreationTimestamp="2025-11-24 17:37:42 +0000 UTC" firstStartedPulling="2025-11-24 17:37:43.972455668 +0000 UTC m=+656.570123470" lastFinishedPulling="2025-11-24 17:37:45.751856193 +0000 UTC m=+658.349523995" observedRunningTime="2025-11-24 17:37:46.245096305 +0000 UTC m=+658.842764107" watchObservedRunningTime="2025-11-24 17:37:46.251348946 +0000 UTC m=+658.849016748" Nov 24 17:37:47 crc kubenswrapper[4808]: I1124 17:37:47.221378 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c" event={"ID":"b9930406-6bd1-4674-a697-e915c209fc08","Type":"ContainerStarted","Data":"67006d163825e073fc48be3937c4e068243a83cd6739ade25342f83b0940839a"} Nov 24 17:37:47 crc kubenswrapper[4808]: I1124 17:37:47.246385 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4x67c" podStartSLOduration=2.352686444 podStartE2EDuration="5.246349843s" podCreationTimestamp="2025-11-24 17:37:42 +0000 UTC" firstStartedPulling="2025-11-24 17:37:44.011478707 +0000 UTC m=+656.609146509" lastFinishedPulling="2025-11-24 17:37:46.905142106 +0000 UTC m=+659.502809908" observedRunningTime="2025-11-24 17:37:47.240056321 +0000 UTC m=+659.837724143" watchObservedRunningTime="2025-11-24 17:37:47.246349843 +0000 UTC m=+659.844017645" Nov 24 17:37:49 crc kubenswrapper[4808]: I1124 17:37:49.238278 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-mf5qh" event={"ID":"125495c2-7b26-44c8-913a-0b98c6c81c98","Type":"ContainerStarted","Data":"a5dffe4dbfdd20046b70b2bb315122986a60b431e8e58a66ec05613d73201406"} Nov 24 17:37:53 crc kubenswrapper[4808]: I1124 17:37:53.128075 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-kz6lk" Nov 24 17:37:53 crc kubenswrapper[4808]: I1124 17:37:53.146412 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-mf5qh" podStartSLOduration=6.319465469 podStartE2EDuration="11.146388818s" podCreationTimestamp="2025-11-24 17:37:42 +0000 UTC" firstStartedPulling="2025-11-24 17:37:43.619479498 +0000 UTC m=+656.217147290" lastFinishedPulling="2025-11-24 17:37:48.446402837 +0000 UTC m=+661.044070639" observedRunningTime="2025-11-24 17:37:49.267756798 +0000 UTC m=+661.865424610" watchObservedRunningTime="2025-11-24 17:37:53.146388818 +0000 UTC m=+665.744056630" Nov 24 17:37:53 crc kubenswrapper[4808]: I1124 17:37:53.548997 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:53 crc kubenswrapper[4808]: I1124 17:37:53.549181 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:53 crc kubenswrapper[4808]: I1124 17:37:53.555369 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:54 crc kubenswrapper[4808]: I1124 17:37:54.270681 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-76fd5f8594-gndtm" Nov 24 17:37:54 crc kubenswrapper[4808]: I1124 17:37:54.331862 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-brt82"] Nov 24 17:38:03 crc kubenswrapper[4808]: I1124 17:38:03.672670 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-rzhnm" Nov 24 17:38:18 crc kubenswrapper[4808]: I1124 17:38:18.619464 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx"] Nov 24 17:38:18 crc kubenswrapper[4808]: I1124 17:38:18.621615 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" Nov 24 17:38:18 crc kubenswrapper[4808]: I1124 17:38:18.624072 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 17:38:18 crc kubenswrapper[4808]: I1124 17:38:18.632345 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx"] Nov 24 17:38:18 crc kubenswrapper[4808]: I1124 17:38:18.715162 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf7p4\" (UniqueName: \"kubernetes.io/projected/9d683a98-fa75-44d9-8d3d-11552b957864-kube-api-access-cf7p4\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx\" (UID: \"9d683a98-fa75-44d9-8d3d-11552b957864\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" Nov 24 17:38:18 crc kubenswrapper[4808]: I1124 17:38:18.715324 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9d683a98-fa75-44d9-8d3d-11552b957864-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx\" (UID: \"9d683a98-fa75-44d9-8d3d-11552b957864\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" Nov 24 17:38:18 crc kubenswrapper[4808]: I1124 17:38:18.715399 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9d683a98-fa75-44d9-8d3d-11552b957864-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx\" (UID: \"9d683a98-fa75-44d9-8d3d-11552b957864\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" Nov 24 17:38:18 crc kubenswrapper[4808]: I1124 17:38:18.816432 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf7p4\" (UniqueName: \"kubernetes.io/projected/9d683a98-fa75-44d9-8d3d-11552b957864-kube-api-access-cf7p4\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx\" (UID: \"9d683a98-fa75-44d9-8d3d-11552b957864\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" Nov 24 17:38:18 crc kubenswrapper[4808]: I1124 17:38:18.816516 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9d683a98-fa75-44d9-8d3d-11552b957864-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx\" (UID: \"9d683a98-fa75-44d9-8d3d-11552b957864\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" Nov 24 17:38:18 crc kubenswrapper[4808]: I1124 17:38:18.816573 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9d683a98-fa75-44d9-8d3d-11552b957864-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx\" (UID: \"9d683a98-fa75-44d9-8d3d-11552b957864\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" Nov 24 17:38:18 crc kubenswrapper[4808]: I1124 17:38:18.817756 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9d683a98-fa75-44d9-8d3d-11552b957864-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx\" (UID: \"9d683a98-fa75-44d9-8d3d-11552b957864\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" Nov 24 17:38:18 crc kubenswrapper[4808]: I1124 17:38:18.818161 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9d683a98-fa75-44d9-8d3d-11552b957864-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx\" (UID: \"9d683a98-fa75-44d9-8d3d-11552b957864\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" Nov 24 17:38:18 crc kubenswrapper[4808]: I1124 17:38:18.841110 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf7p4\" (UniqueName: \"kubernetes.io/projected/9d683a98-fa75-44d9-8d3d-11552b957864-kube-api-access-cf7p4\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx\" (UID: \"9d683a98-fa75-44d9-8d3d-11552b957864\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" Nov 24 17:38:18 crc kubenswrapper[4808]: I1124 17:38:18.951830 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" Nov 24 17:38:19 crc kubenswrapper[4808]: I1124 17:38:19.375136 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-brt82" podUID="e87a6b3d-2bbf-49a1-9209-a1327c4bd30d" containerName="console" containerID="cri-o://380916c934587ca527f21d29d3e3721f2b54c362661e7e73e5643f4806db8801" gracePeriod=15 Nov 24 17:38:19 crc kubenswrapper[4808]: I1124 17:38:19.385350 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx"] Nov 24 17:38:19 crc kubenswrapper[4808]: I1124 17:38:19.430621 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" event={"ID":"9d683a98-fa75-44d9-8d3d-11552b957864","Type":"ContainerStarted","Data":"83f690fc5da36396518de1c2d9dc1a566cc2087e6cc8b17484ce40056c39b9a5"} Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.009174 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-brt82_e87a6b3d-2bbf-49a1-9209-a1327c4bd30d/console/0.log" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.009276 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.035884 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-oauth-config\") pod \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.035974 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-trusted-ca-bundle\") pod \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.036010 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-service-ca\") pod \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.036120 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-oauth-serving-cert\") pod \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.036171 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-config\") pod \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.036202 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-serving-cert\") pod \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.036244 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x8nr\" (UniqueName: \"kubernetes.io/projected/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-kube-api-access-4x8nr\") pod \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\" (UID: \"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d\") " Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.043489 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-service-ca" (OuterVolumeSpecName: "service-ca") pod "e87a6b3d-2bbf-49a1-9209-a1327c4bd30d" (UID: "e87a6b3d-2bbf-49a1-9209-a1327c4bd30d"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.044905 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-config" (OuterVolumeSpecName: "console-config") pod "e87a6b3d-2bbf-49a1-9209-a1327c4bd30d" (UID: "e87a6b3d-2bbf-49a1-9209-a1327c4bd30d"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.045080 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "e87a6b3d-2bbf-49a1-9209-a1327c4bd30d" (UID: "e87a6b3d-2bbf-49a1-9209-a1327c4bd30d"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.045122 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "e87a6b3d-2bbf-49a1-9209-a1327c4bd30d" (UID: "e87a6b3d-2bbf-49a1-9209-a1327c4bd30d"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.050379 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "e87a6b3d-2bbf-49a1-9209-a1327c4bd30d" (UID: "e87a6b3d-2bbf-49a1-9209-a1327c4bd30d"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.054275 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "e87a6b3d-2bbf-49a1-9209-a1327c4bd30d" (UID: "e87a6b3d-2bbf-49a1-9209-a1327c4bd30d"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.054471 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-kube-api-access-4x8nr" (OuterVolumeSpecName: "kube-api-access-4x8nr") pod "e87a6b3d-2bbf-49a1-9209-a1327c4bd30d" (UID: "e87a6b3d-2bbf-49a1-9209-a1327c4bd30d"). InnerVolumeSpecName "kube-api-access-4x8nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.138314 4808 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.138853 4808 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.138863 4808 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.138875 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x8nr\" (UniqueName: \"kubernetes.io/projected/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-kube-api-access-4x8nr\") on node \"crc\" DevicePath \"\"" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.138887 4808 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.138896 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.138909 4808 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.439940 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-brt82_e87a6b3d-2bbf-49a1-9209-a1327c4bd30d/console/0.log" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.440006 4808 generic.go:334] "Generic (PLEG): container finished" podID="e87a6b3d-2bbf-49a1-9209-a1327c4bd30d" containerID="380916c934587ca527f21d29d3e3721f2b54c362661e7e73e5643f4806db8801" exitCode=2 Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.440091 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-brt82" event={"ID":"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d","Type":"ContainerDied","Data":"380916c934587ca527f21d29d3e3721f2b54c362661e7e73e5643f4806db8801"} Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.440176 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-brt82" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.440205 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-brt82" event={"ID":"e87a6b3d-2bbf-49a1-9209-a1327c4bd30d","Type":"ContainerDied","Data":"b747ed7262c0d73c9e8deb41ab48fcd9b1133c1d57bd14c438347f06773b9199"} Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.440231 4808 scope.go:117] "RemoveContainer" containerID="380916c934587ca527f21d29d3e3721f2b54c362661e7e73e5643f4806db8801" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.444136 4808 generic.go:334] "Generic (PLEG): container finished" podID="9d683a98-fa75-44d9-8d3d-11552b957864" containerID="111888a4f849d21fd195e5bd3c0ad9170afa93710d3b97d266ed4a66bccf5d75" exitCode=0 Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.444188 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" event={"ID":"9d683a98-fa75-44d9-8d3d-11552b957864","Type":"ContainerDied","Data":"111888a4f849d21fd195e5bd3c0ad9170afa93710d3b97d266ed4a66bccf5d75"} Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.466397 4808 scope.go:117] "RemoveContainer" containerID="380916c934587ca527f21d29d3e3721f2b54c362661e7e73e5643f4806db8801" Nov 24 17:38:20 crc kubenswrapper[4808]: E1124 17:38:20.468925 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"380916c934587ca527f21d29d3e3721f2b54c362661e7e73e5643f4806db8801\": container with ID starting with 380916c934587ca527f21d29d3e3721f2b54c362661e7e73e5643f4806db8801 not found: ID does not exist" containerID="380916c934587ca527f21d29d3e3721f2b54c362661e7e73e5643f4806db8801" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.468958 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"380916c934587ca527f21d29d3e3721f2b54c362661e7e73e5643f4806db8801"} err="failed to get container status \"380916c934587ca527f21d29d3e3721f2b54c362661e7e73e5643f4806db8801\": rpc error: code = NotFound desc = could not find container \"380916c934587ca527f21d29d3e3721f2b54c362661e7e73e5643f4806db8801\": container with ID starting with 380916c934587ca527f21d29d3e3721f2b54c362661e7e73e5643f4806db8801 not found: ID does not exist" Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.481943 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-brt82"] Nov 24 17:38:20 crc kubenswrapper[4808]: I1124 17:38:20.486841 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-brt82"] Nov 24 17:38:22 crc kubenswrapper[4808]: I1124 17:38:22.356486 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e87a6b3d-2bbf-49a1-9209-a1327c4bd30d" path="/var/lib/kubelet/pods/e87a6b3d-2bbf-49a1-9209-a1327c4bd30d/volumes" Nov 24 17:38:22 crc kubenswrapper[4808]: I1124 17:38:22.461951 4808 generic.go:334] "Generic (PLEG): container finished" podID="9d683a98-fa75-44d9-8d3d-11552b957864" containerID="5b1194b9d80c13a278c5237199b7b21efbdb20784f0533b45141f3094fcdfde9" exitCode=0 Nov 24 17:38:22 crc kubenswrapper[4808]: I1124 17:38:22.461996 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" event={"ID":"9d683a98-fa75-44d9-8d3d-11552b957864","Type":"ContainerDied","Data":"5b1194b9d80c13a278c5237199b7b21efbdb20784f0533b45141f3094fcdfde9"} Nov 24 17:38:23 crc kubenswrapper[4808]: I1124 17:38:23.470936 4808 generic.go:334] "Generic (PLEG): container finished" podID="9d683a98-fa75-44d9-8d3d-11552b957864" containerID="404bbb08a1482d810ace7bc90a5ca9a37fd51df9cfcf6db3899b5ce1c647a017" exitCode=0 Nov 24 17:38:23 crc kubenswrapper[4808]: I1124 17:38:23.471005 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" event={"ID":"9d683a98-fa75-44d9-8d3d-11552b957864","Type":"ContainerDied","Data":"404bbb08a1482d810ace7bc90a5ca9a37fd51df9cfcf6db3899b5ce1c647a017"} Nov 24 17:38:24 crc kubenswrapper[4808]: I1124 17:38:24.717162 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" Nov 24 17:38:24 crc kubenswrapper[4808]: I1124 17:38:24.808834 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9d683a98-fa75-44d9-8d3d-11552b957864-util\") pod \"9d683a98-fa75-44d9-8d3d-11552b957864\" (UID: \"9d683a98-fa75-44d9-8d3d-11552b957864\") " Nov 24 17:38:24 crc kubenswrapper[4808]: I1124 17:38:24.808984 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9d683a98-fa75-44d9-8d3d-11552b957864-bundle\") pod \"9d683a98-fa75-44d9-8d3d-11552b957864\" (UID: \"9d683a98-fa75-44d9-8d3d-11552b957864\") " Nov 24 17:38:24 crc kubenswrapper[4808]: I1124 17:38:24.809155 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf7p4\" (UniqueName: \"kubernetes.io/projected/9d683a98-fa75-44d9-8d3d-11552b957864-kube-api-access-cf7p4\") pod \"9d683a98-fa75-44d9-8d3d-11552b957864\" (UID: \"9d683a98-fa75-44d9-8d3d-11552b957864\") " Nov 24 17:38:24 crc kubenswrapper[4808]: I1124 17:38:24.810977 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d683a98-fa75-44d9-8d3d-11552b957864-bundle" (OuterVolumeSpecName: "bundle") pod "9d683a98-fa75-44d9-8d3d-11552b957864" (UID: "9d683a98-fa75-44d9-8d3d-11552b957864"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:38:24 crc kubenswrapper[4808]: I1124 17:38:24.815330 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d683a98-fa75-44d9-8d3d-11552b957864-kube-api-access-cf7p4" (OuterVolumeSpecName: "kube-api-access-cf7p4") pod "9d683a98-fa75-44d9-8d3d-11552b957864" (UID: "9d683a98-fa75-44d9-8d3d-11552b957864"). InnerVolumeSpecName "kube-api-access-cf7p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:38:24 crc kubenswrapper[4808]: I1124 17:38:24.827555 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d683a98-fa75-44d9-8d3d-11552b957864-util" (OuterVolumeSpecName: "util") pod "9d683a98-fa75-44d9-8d3d-11552b957864" (UID: "9d683a98-fa75-44d9-8d3d-11552b957864"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:38:24 crc kubenswrapper[4808]: I1124 17:38:24.910766 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf7p4\" (UniqueName: \"kubernetes.io/projected/9d683a98-fa75-44d9-8d3d-11552b957864-kube-api-access-cf7p4\") on node \"crc\" DevicePath \"\"" Nov 24 17:38:24 crc kubenswrapper[4808]: I1124 17:38:24.910818 4808 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9d683a98-fa75-44d9-8d3d-11552b957864-util\") on node \"crc\" DevicePath \"\"" Nov 24 17:38:24 crc kubenswrapper[4808]: I1124 17:38:24.910831 4808 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9d683a98-fa75-44d9-8d3d-11552b957864-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:38:25 crc kubenswrapper[4808]: I1124 17:38:25.485911 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" event={"ID":"9d683a98-fa75-44d9-8d3d-11552b957864","Type":"ContainerDied","Data":"83f690fc5da36396518de1c2d9dc1a566cc2087e6cc8b17484ce40056c39b9a5"} Nov 24 17:38:25 crc kubenswrapper[4808]: I1124 17:38:25.485953 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx" Nov 24 17:38:25 crc kubenswrapper[4808]: I1124 17:38:25.485956 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83f690fc5da36396518de1c2d9dc1a566cc2087e6cc8b17484ce40056c39b9a5" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.229554 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-56c4766567-nztcn"] Nov 24 17:38:37 crc kubenswrapper[4808]: E1124 17:38:37.230686 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d683a98-fa75-44d9-8d3d-11552b957864" containerName="pull" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.230705 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d683a98-fa75-44d9-8d3d-11552b957864" containerName="pull" Nov 24 17:38:37 crc kubenswrapper[4808]: E1124 17:38:37.230714 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e87a6b3d-2bbf-49a1-9209-a1327c4bd30d" containerName="console" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.230721 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e87a6b3d-2bbf-49a1-9209-a1327c4bd30d" containerName="console" Nov 24 17:38:37 crc kubenswrapper[4808]: E1124 17:38:37.230737 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d683a98-fa75-44d9-8d3d-11552b957864" containerName="extract" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.230743 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d683a98-fa75-44d9-8d3d-11552b957864" containerName="extract" Nov 24 17:38:37 crc kubenswrapper[4808]: E1124 17:38:37.230752 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d683a98-fa75-44d9-8d3d-11552b957864" containerName="util" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.230759 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d683a98-fa75-44d9-8d3d-11552b957864" containerName="util" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.230894 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d683a98-fa75-44d9-8d3d-11552b957864" containerName="extract" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.230903 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e87a6b3d-2bbf-49a1-9209-a1327c4bd30d" containerName="console" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.236400 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.248188 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.248498 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-jdvlb" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.248780 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.249425 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.253193 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.266102 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-56c4766567-nztcn"] Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.292226 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8dc3fedf-199a-4446-a18a-3dd0f6818f8d-webhook-cert\") pod \"metallb-operator-controller-manager-56c4766567-nztcn\" (UID: \"8dc3fedf-199a-4446-a18a-3dd0f6818f8d\") " pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.292355 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8dc3fedf-199a-4446-a18a-3dd0f6818f8d-apiservice-cert\") pod \"metallb-operator-controller-manager-56c4766567-nztcn\" (UID: \"8dc3fedf-199a-4446-a18a-3dd0f6818f8d\") " pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.292395 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2g2m\" (UniqueName: \"kubernetes.io/projected/8dc3fedf-199a-4446-a18a-3dd0f6818f8d-kube-api-access-n2g2m\") pod \"metallb-operator-controller-manager-56c4766567-nztcn\" (UID: \"8dc3fedf-199a-4446-a18a-3dd0f6818f8d\") " pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.393863 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8dc3fedf-199a-4446-a18a-3dd0f6818f8d-webhook-cert\") pod \"metallb-operator-controller-manager-56c4766567-nztcn\" (UID: \"8dc3fedf-199a-4446-a18a-3dd0f6818f8d\") " pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.393987 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8dc3fedf-199a-4446-a18a-3dd0f6818f8d-apiservice-cert\") pod \"metallb-operator-controller-manager-56c4766567-nztcn\" (UID: \"8dc3fedf-199a-4446-a18a-3dd0f6818f8d\") " pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.394133 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2g2m\" (UniqueName: \"kubernetes.io/projected/8dc3fedf-199a-4446-a18a-3dd0f6818f8d-kube-api-access-n2g2m\") pod \"metallb-operator-controller-manager-56c4766567-nztcn\" (UID: \"8dc3fedf-199a-4446-a18a-3dd0f6818f8d\") " pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.404779 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8dc3fedf-199a-4446-a18a-3dd0f6818f8d-apiservice-cert\") pod \"metallb-operator-controller-manager-56c4766567-nztcn\" (UID: \"8dc3fedf-199a-4446-a18a-3dd0f6818f8d\") " pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.405803 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8dc3fedf-199a-4446-a18a-3dd0f6818f8d-webhook-cert\") pod \"metallb-operator-controller-manager-56c4766567-nztcn\" (UID: \"8dc3fedf-199a-4446-a18a-3dd0f6818f8d\") " pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.436141 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2g2m\" (UniqueName: \"kubernetes.io/projected/8dc3fedf-199a-4446-a18a-3dd0f6818f8d-kube-api-access-n2g2m\") pod \"metallb-operator-controller-manager-56c4766567-nztcn\" (UID: \"8dc3fedf-199a-4446-a18a-3dd0f6818f8d\") " pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.566369 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.575511 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5"] Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.576785 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.583989 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.584043 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.584241 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-zvhsh" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.594892 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5"] Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.597811 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/da67a1e6-9fd1-4b27-b965-c907e51b6ce4-webhook-cert\") pod \"metallb-operator-webhook-server-7c6ddcbc67-pn9v5\" (UID: \"da67a1e6-9fd1-4b27-b965-c907e51b6ce4\") " pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.597894 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/da67a1e6-9fd1-4b27-b965-c907e51b6ce4-apiservice-cert\") pod \"metallb-operator-webhook-server-7c6ddcbc67-pn9v5\" (UID: \"da67a1e6-9fd1-4b27-b965-c907e51b6ce4\") " pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.597975 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzllx\" (UniqueName: \"kubernetes.io/projected/da67a1e6-9fd1-4b27-b965-c907e51b6ce4-kube-api-access-wzllx\") pod \"metallb-operator-webhook-server-7c6ddcbc67-pn9v5\" (UID: \"da67a1e6-9fd1-4b27-b965-c907e51b6ce4\") " pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.699153 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/da67a1e6-9fd1-4b27-b965-c907e51b6ce4-webhook-cert\") pod \"metallb-operator-webhook-server-7c6ddcbc67-pn9v5\" (UID: \"da67a1e6-9fd1-4b27-b965-c907e51b6ce4\") " pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.699220 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/da67a1e6-9fd1-4b27-b965-c907e51b6ce4-apiservice-cert\") pod \"metallb-operator-webhook-server-7c6ddcbc67-pn9v5\" (UID: \"da67a1e6-9fd1-4b27-b965-c907e51b6ce4\") " pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.699297 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzllx\" (UniqueName: \"kubernetes.io/projected/da67a1e6-9fd1-4b27-b965-c907e51b6ce4-kube-api-access-wzllx\") pod \"metallb-operator-webhook-server-7c6ddcbc67-pn9v5\" (UID: \"da67a1e6-9fd1-4b27-b965-c907e51b6ce4\") " pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.704588 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/da67a1e6-9fd1-4b27-b965-c907e51b6ce4-webhook-cert\") pod \"metallb-operator-webhook-server-7c6ddcbc67-pn9v5\" (UID: \"da67a1e6-9fd1-4b27-b965-c907e51b6ce4\") " pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.706880 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/da67a1e6-9fd1-4b27-b965-c907e51b6ce4-apiservice-cert\") pod \"metallb-operator-webhook-server-7c6ddcbc67-pn9v5\" (UID: \"da67a1e6-9fd1-4b27-b965-c907e51b6ce4\") " pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.732194 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzllx\" (UniqueName: \"kubernetes.io/projected/da67a1e6-9fd1-4b27-b965-c907e51b6ce4-kube-api-access-wzllx\") pod \"metallb-operator-webhook-server-7c6ddcbc67-pn9v5\" (UID: \"da67a1e6-9fd1-4b27-b965-c907e51b6ce4\") " pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.954170 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-56c4766567-nztcn"] Nov 24 17:38:37 crc kubenswrapper[4808]: I1124 17:38:37.963050 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" Nov 24 17:38:38 crc kubenswrapper[4808]: I1124 17:38:38.232705 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5"] Nov 24 17:38:38 crc kubenswrapper[4808]: W1124 17:38:38.241820 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda67a1e6_9fd1_4b27_b965_c907e51b6ce4.slice/crio-dc8d74340e6166827a26ccf6420fdb99ad52f5beab6ffa69d5bb704dbf1cd9d2 WatchSource:0}: Error finding container dc8d74340e6166827a26ccf6420fdb99ad52f5beab6ffa69d5bb704dbf1cd9d2: Status 404 returned error can't find the container with id dc8d74340e6166827a26ccf6420fdb99ad52f5beab6ffa69d5bb704dbf1cd9d2 Nov 24 17:38:38 crc kubenswrapper[4808]: I1124 17:38:38.567535 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" event={"ID":"da67a1e6-9fd1-4b27-b965-c907e51b6ce4","Type":"ContainerStarted","Data":"dc8d74340e6166827a26ccf6420fdb99ad52f5beab6ffa69d5bb704dbf1cd9d2"} Nov 24 17:38:38 crc kubenswrapper[4808]: I1124 17:38:38.569354 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" event={"ID":"8dc3fedf-199a-4446-a18a-3dd0f6818f8d","Type":"ContainerStarted","Data":"8a7d9acf3fc55c67a81903ec723c5a759cc6b3942b6760cfafe20b7a37d44fa4"} Nov 24 17:38:43 crc kubenswrapper[4808]: I1124 17:38:43.610630 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" event={"ID":"8dc3fedf-199a-4446-a18a-3dd0f6818f8d","Type":"ContainerStarted","Data":"77264aea65924e8cda066810d0eed30df8fe418f9bc9300f85bb7d73ee49a004"} Nov 24 17:38:43 crc kubenswrapper[4808]: I1124 17:38:43.611552 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" Nov 24 17:38:43 crc kubenswrapper[4808]: I1124 17:38:43.613332 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" event={"ID":"da67a1e6-9fd1-4b27-b965-c907e51b6ce4","Type":"ContainerStarted","Data":"63b3f1e69d18af1186cec5f3048e4ebbd77a6aa69792f449e6fa6a60bc89bb2b"} Nov 24 17:38:43 crc kubenswrapper[4808]: I1124 17:38:43.613552 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" Nov 24 17:38:43 crc kubenswrapper[4808]: I1124 17:38:43.659442 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" podStartSLOduration=1.8604476540000001 podStartE2EDuration="6.659412407s" podCreationTimestamp="2025-11-24 17:38:37 +0000 UTC" firstStartedPulling="2025-11-24 17:38:37.971722557 +0000 UTC m=+710.569390359" lastFinishedPulling="2025-11-24 17:38:42.77068731 +0000 UTC m=+715.368355112" observedRunningTime="2025-11-24 17:38:43.638694829 +0000 UTC m=+716.236362631" watchObservedRunningTime="2025-11-24 17:38:43.659412407 +0000 UTC m=+716.257080199" Nov 24 17:38:43 crc kubenswrapper[4808]: I1124 17:38:43.660348 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" podStartSLOduration=2.117096483 podStartE2EDuration="6.660340072s" podCreationTimestamp="2025-11-24 17:38:37 +0000 UTC" firstStartedPulling="2025-11-24 17:38:38.245674611 +0000 UTC m=+710.843342413" lastFinishedPulling="2025-11-24 17:38:42.7889182 +0000 UTC m=+715.386586002" observedRunningTime="2025-11-24 17:38:43.656827016 +0000 UTC m=+716.254494818" watchObservedRunningTime="2025-11-24 17:38:43.660340072 +0000 UTC m=+716.258007874" Nov 24 17:38:57 crc kubenswrapper[4808]: I1124 17:38:57.969451 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7c6ddcbc67-pn9v5" Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.209657 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vwchc"] Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.210835 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" podUID="31954b89-9642-4b55-b21e-79882b30b726" containerName="controller-manager" containerID="cri-o://f37f56bdf2b293663f28edac5f500a763a42c2a2a0edae8a045345ddd2704f6b" gracePeriod=30 Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.311820 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp"] Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.312598 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" podUID="50f0b611-6832-4a5d-8afb-13d5724e8019" containerName="route-controller-manager" containerID="cri-o://1a216ed389533603c08fb010bd19f359db30e13e36256973b9cce788bd7d8a80" gracePeriod=30 Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.736141 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.825169 4808 generic.go:334] "Generic (PLEG): container finished" podID="31954b89-9642-4b55-b21e-79882b30b726" containerID="f37f56bdf2b293663f28edac5f500a763a42c2a2a0edae8a045345ddd2704f6b" exitCode=0 Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.825257 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.825287 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" event={"ID":"31954b89-9642-4b55-b21e-79882b30b726","Type":"ContainerDied","Data":"f37f56bdf2b293663f28edac5f500a763a42c2a2a0edae8a045345ddd2704f6b"} Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.825364 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vwchc" event={"ID":"31954b89-9642-4b55-b21e-79882b30b726","Type":"ContainerDied","Data":"aa8eb98c2ebd24ee01f789edf13676f8b49c05fcea68b911bb71dbedb78371fb"} Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.825393 4808 scope.go:117] "RemoveContainer" containerID="f37f56bdf2b293663f28edac5f500a763a42c2a2a0edae8a045345ddd2704f6b" Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.827111 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-client-ca\") pod \"31954b89-9642-4b55-b21e-79882b30b726\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.827145 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-proxy-ca-bundles\") pod \"31954b89-9642-4b55-b21e-79882b30b726\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.827311 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31954b89-9642-4b55-b21e-79882b30b726-serving-cert\") pod \"31954b89-9642-4b55-b21e-79882b30b726\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.827334 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-config\") pod \"31954b89-9642-4b55-b21e-79882b30b726\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.827366 4808 generic.go:334] "Generic (PLEG): container finished" podID="50f0b611-6832-4a5d-8afb-13d5724e8019" containerID="1a216ed389533603c08fb010bd19f359db30e13e36256973b9cce788bd7d8a80" exitCode=0 Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.827394 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" event={"ID":"50f0b611-6832-4a5d-8afb-13d5724e8019","Type":"ContainerDied","Data":"1a216ed389533603c08fb010bd19f359db30e13e36256973b9cce788bd7d8a80"} Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.827392 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mdcx\" (UniqueName: \"kubernetes.io/projected/31954b89-9642-4b55-b21e-79882b30b726-kube-api-access-7mdcx\") pod \"31954b89-9642-4b55-b21e-79882b30b726\" (UID: \"31954b89-9642-4b55-b21e-79882b30b726\") " Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.828456 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-client-ca" (OuterVolumeSpecName: "client-ca") pod "31954b89-9642-4b55-b21e-79882b30b726" (UID: "31954b89-9642-4b55-b21e-79882b30b726"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.829408 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "31954b89-9642-4b55-b21e-79882b30b726" (UID: "31954b89-9642-4b55-b21e-79882b30b726"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.829524 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-config" (OuterVolumeSpecName: "config") pod "31954b89-9642-4b55-b21e-79882b30b726" (UID: "31954b89-9642-4b55-b21e-79882b30b726"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.835636 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31954b89-9642-4b55-b21e-79882b30b726-kube-api-access-7mdcx" (OuterVolumeSpecName: "kube-api-access-7mdcx") pod "31954b89-9642-4b55-b21e-79882b30b726" (UID: "31954b89-9642-4b55-b21e-79882b30b726"). InnerVolumeSpecName "kube-api-access-7mdcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.836446 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31954b89-9642-4b55-b21e-79882b30b726-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "31954b89-9642-4b55-b21e-79882b30b726" (UID: "31954b89-9642-4b55-b21e-79882b30b726"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.881597 4808 scope.go:117] "RemoveContainer" containerID="f37f56bdf2b293663f28edac5f500a763a42c2a2a0edae8a045345ddd2704f6b" Nov 24 17:39:12 crc kubenswrapper[4808]: E1124 17:39:12.882331 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f37f56bdf2b293663f28edac5f500a763a42c2a2a0edae8a045345ddd2704f6b\": container with ID starting with f37f56bdf2b293663f28edac5f500a763a42c2a2a0edae8a045345ddd2704f6b not found: ID does not exist" containerID="f37f56bdf2b293663f28edac5f500a763a42c2a2a0edae8a045345ddd2704f6b" Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.882405 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f37f56bdf2b293663f28edac5f500a763a42c2a2a0edae8a045345ddd2704f6b"} err="failed to get container status \"f37f56bdf2b293663f28edac5f500a763a42c2a2a0edae8a045345ddd2704f6b\": rpc error: code = NotFound desc = could not find container \"f37f56bdf2b293663f28edac5f500a763a42c2a2a0edae8a045345ddd2704f6b\": container with ID starting with f37f56bdf2b293663f28edac5f500a763a42c2a2a0edae8a045345ddd2704f6b not found: ID does not exist" Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.929313 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31954b89-9642-4b55-b21e-79882b30b726-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.930076 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.930135 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mdcx\" (UniqueName: \"kubernetes.io/projected/31954b89-9642-4b55-b21e-79882b30b726-kube-api-access-7mdcx\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.930153 4808 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:12 crc kubenswrapper[4808]: I1124 17:39:12.930170 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/31954b89-9642-4b55-b21e-79882b30b726-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.161267 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vwchc"] Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.165336 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vwchc"] Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.205064 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.335961 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdlbv\" (UniqueName: \"kubernetes.io/projected/50f0b611-6832-4a5d-8afb-13d5724e8019-kube-api-access-gdlbv\") pod \"50f0b611-6832-4a5d-8afb-13d5724e8019\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.336147 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50f0b611-6832-4a5d-8afb-13d5724e8019-client-ca\") pod \"50f0b611-6832-4a5d-8afb-13d5724e8019\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.336175 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50f0b611-6832-4a5d-8afb-13d5724e8019-serving-cert\") pod \"50f0b611-6832-4a5d-8afb-13d5724e8019\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.336244 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50f0b611-6832-4a5d-8afb-13d5724e8019-config\") pod \"50f0b611-6832-4a5d-8afb-13d5724e8019\" (UID: \"50f0b611-6832-4a5d-8afb-13d5724e8019\") " Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.337670 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50f0b611-6832-4a5d-8afb-13d5724e8019-config" (OuterVolumeSpecName: "config") pod "50f0b611-6832-4a5d-8afb-13d5724e8019" (UID: "50f0b611-6832-4a5d-8afb-13d5724e8019"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.337984 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50f0b611-6832-4a5d-8afb-13d5724e8019-client-ca" (OuterVolumeSpecName: "client-ca") pod "50f0b611-6832-4a5d-8afb-13d5724e8019" (UID: "50f0b611-6832-4a5d-8afb-13d5724e8019"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.338266 4808 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50f0b611-6832-4a5d-8afb-13d5724e8019-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.338297 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50f0b611-6832-4a5d-8afb-13d5724e8019-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.341995 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f0b611-6832-4a5d-8afb-13d5724e8019-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "50f0b611-6832-4a5d-8afb-13d5724e8019" (UID: "50f0b611-6832-4a5d-8afb-13d5724e8019"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.342048 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50f0b611-6832-4a5d-8afb-13d5724e8019-kube-api-access-gdlbv" (OuterVolumeSpecName: "kube-api-access-gdlbv") pod "50f0b611-6832-4a5d-8afb-13d5724e8019" (UID: "50f0b611-6832-4a5d-8afb-13d5724e8019"). InnerVolumeSpecName "kube-api-access-gdlbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.375805 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb"] Nov 24 17:39:13 crc kubenswrapper[4808]: E1124 17:39:13.376142 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31954b89-9642-4b55-b21e-79882b30b726" containerName="controller-manager" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.376163 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="31954b89-9642-4b55-b21e-79882b30b726" containerName="controller-manager" Nov 24 17:39:13 crc kubenswrapper[4808]: E1124 17:39:13.376173 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50f0b611-6832-4a5d-8afb-13d5724e8019" containerName="route-controller-manager" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.376181 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="50f0b611-6832-4a5d-8afb-13d5724e8019" containerName="route-controller-manager" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.376324 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="50f0b611-6832-4a5d-8afb-13d5724e8019" containerName="route-controller-manager" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.376347 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="31954b89-9642-4b55-b21e-79882b30b726" containerName="controller-manager" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.376863 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.387852 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb"] Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.439096 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4efa838-4f42-41e3-b12c-4574585bd8f4-serving-cert\") pod \"route-controller-manager-688c64f9c6-dz2zb\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.439168 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcm9m\" (UniqueName: \"kubernetes.io/projected/c4efa838-4f42-41e3-b12c-4574585bd8f4-kube-api-access-fcm9m\") pod \"route-controller-manager-688c64f9c6-dz2zb\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.439663 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4efa838-4f42-41e3-b12c-4574585bd8f4-config\") pod \"route-controller-manager-688c64f9c6-dz2zb\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.439760 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4efa838-4f42-41e3-b12c-4574585bd8f4-client-ca\") pod \"route-controller-manager-688c64f9c6-dz2zb\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.439834 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdlbv\" (UniqueName: \"kubernetes.io/projected/50f0b611-6832-4a5d-8afb-13d5724e8019-kube-api-access-gdlbv\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.439849 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50f0b611-6832-4a5d-8afb-13d5724e8019-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.541715 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4efa838-4f42-41e3-b12c-4574585bd8f4-config\") pod \"route-controller-manager-688c64f9c6-dz2zb\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.541796 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4efa838-4f42-41e3-b12c-4574585bd8f4-client-ca\") pod \"route-controller-manager-688c64f9c6-dz2zb\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.541816 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4efa838-4f42-41e3-b12c-4574585bd8f4-serving-cert\") pod \"route-controller-manager-688c64f9c6-dz2zb\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.541832 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcm9m\" (UniqueName: \"kubernetes.io/projected/c4efa838-4f42-41e3-b12c-4574585bd8f4-kube-api-access-fcm9m\") pod \"route-controller-manager-688c64f9c6-dz2zb\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.543694 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4efa838-4f42-41e3-b12c-4574585bd8f4-config\") pod \"route-controller-manager-688c64f9c6-dz2zb\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.544298 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4efa838-4f42-41e3-b12c-4574585bd8f4-client-ca\") pod \"route-controller-manager-688c64f9c6-dz2zb\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.550622 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4efa838-4f42-41e3-b12c-4574585bd8f4-serving-cert\") pod \"route-controller-manager-688c64f9c6-dz2zb\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.561255 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcm9m\" (UniqueName: \"kubernetes.io/projected/c4efa838-4f42-41e3-b12c-4574585bd8f4-kube-api-access-fcm9m\") pod \"route-controller-manager-688c64f9c6-dz2zb\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.703815 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.839661 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" event={"ID":"50f0b611-6832-4a5d-8afb-13d5724e8019","Type":"ContainerDied","Data":"c9e8f19112e3df39d36bc1fa35a9a22f6691de6466e4d480b1d90d7133e205b2"} Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.840189 4808 scope.go:117] "RemoveContainer" containerID="1a216ed389533603c08fb010bd19f359db30e13e36256973b9cce788bd7d8a80" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.839742 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp" Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.884395 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp"] Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.887362 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-x5mbp"] Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.963254 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb"] Nov 24 17:39:13 crc kubenswrapper[4808]: I1124 17:39:13.992832 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb"] Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.355405 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31954b89-9642-4b55-b21e-79882b30b726" path="/var/lib/kubelet/pods/31954b89-9642-4b55-b21e-79882b30b726/volumes" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.356363 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50f0b611-6832-4a5d-8afb-13d5724e8019" path="/var/lib/kubelet/pods/50f0b611-6832-4a5d-8afb-13d5724e8019/volumes" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.374984 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-66bc449dc9-gnnjk"] Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.376102 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.381252 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.381527 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.381965 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.382837 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.383963 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.384180 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.403693 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.419738 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66bc449dc9-gnnjk"] Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.452994 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36a6c0d1-772b-4ea4-a425-3a59b861f297-serving-cert\") pod \"controller-manager-66bc449dc9-gnnjk\" (UID: \"36a6c0d1-772b-4ea4-a425-3a59b861f297\") " pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.453087 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36a6c0d1-772b-4ea4-a425-3a59b861f297-proxy-ca-bundles\") pod \"controller-manager-66bc449dc9-gnnjk\" (UID: \"36a6c0d1-772b-4ea4-a425-3a59b861f297\") " pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.453114 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5hkf\" (UniqueName: \"kubernetes.io/projected/36a6c0d1-772b-4ea4-a425-3a59b861f297-kube-api-access-m5hkf\") pod \"controller-manager-66bc449dc9-gnnjk\" (UID: \"36a6c0d1-772b-4ea4-a425-3a59b861f297\") " pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.453147 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36a6c0d1-772b-4ea4-a425-3a59b861f297-client-ca\") pod \"controller-manager-66bc449dc9-gnnjk\" (UID: \"36a6c0d1-772b-4ea4-a425-3a59b861f297\") " pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.453223 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36a6c0d1-772b-4ea4-a425-3a59b861f297-config\") pod \"controller-manager-66bc449dc9-gnnjk\" (UID: \"36a6c0d1-772b-4ea4-a425-3a59b861f297\") " pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.554552 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36a6c0d1-772b-4ea4-a425-3a59b861f297-config\") pod \"controller-manager-66bc449dc9-gnnjk\" (UID: \"36a6c0d1-772b-4ea4-a425-3a59b861f297\") " pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.554628 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36a6c0d1-772b-4ea4-a425-3a59b861f297-serving-cert\") pod \"controller-manager-66bc449dc9-gnnjk\" (UID: \"36a6c0d1-772b-4ea4-a425-3a59b861f297\") " pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.554664 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36a6c0d1-772b-4ea4-a425-3a59b861f297-proxy-ca-bundles\") pod \"controller-manager-66bc449dc9-gnnjk\" (UID: \"36a6c0d1-772b-4ea4-a425-3a59b861f297\") " pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.554680 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5hkf\" (UniqueName: \"kubernetes.io/projected/36a6c0d1-772b-4ea4-a425-3a59b861f297-kube-api-access-m5hkf\") pod \"controller-manager-66bc449dc9-gnnjk\" (UID: \"36a6c0d1-772b-4ea4-a425-3a59b861f297\") " pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.554707 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36a6c0d1-772b-4ea4-a425-3a59b861f297-client-ca\") pod \"controller-manager-66bc449dc9-gnnjk\" (UID: \"36a6c0d1-772b-4ea4-a425-3a59b861f297\") " pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.556449 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36a6c0d1-772b-4ea4-a425-3a59b861f297-client-ca\") pod \"controller-manager-66bc449dc9-gnnjk\" (UID: \"36a6c0d1-772b-4ea4-a425-3a59b861f297\") " pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.556782 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36a6c0d1-772b-4ea4-a425-3a59b861f297-config\") pod \"controller-manager-66bc449dc9-gnnjk\" (UID: \"36a6c0d1-772b-4ea4-a425-3a59b861f297\") " pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.557901 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36a6c0d1-772b-4ea4-a425-3a59b861f297-proxy-ca-bundles\") pod \"controller-manager-66bc449dc9-gnnjk\" (UID: \"36a6c0d1-772b-4ea4-a425-3a59b861f297\") " pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.563803 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36a6c0d1-772b-4ea4-a425-3a59b861f297-serving-cert\") pod \"controller-manager-66bc449dc9-gnnjk\" (UID: \"36a6c0d1-772b-4ea4-a425-3a59b861f297\") " pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.585864 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5hkf\" (UniqueName: \"kubernetes.io/projected/36a6c0d1-772b-4ea4-a425-3a59b861f297-kube-api-access-m5hkf\") pod \"controller-manager-66bc449dc9-gnnjk\" (UID: \"36a6c0d1-772b-4ea4-a425-3a59b861f297\") " pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.708659 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.851768 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" event={"ID":"c4efa838-4f42-41e3-b12c-4574585bd8f4","Type":"ContainerStarted","Data":"24055d9a4fe4ecdd8a9c53f05703e7a5f45b0adad3c69d4a6764d0d2323f230d"} Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.852313 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" event={"ID":"c4efa838-4f42-41e3-b12c-4574585bd8f4","Type":"ContainerStarted","Data":"a8bc4aa9005c29b6c79dfb2baed3d21938d47ddd52189ce025768db63962cf58"} Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.852180 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" podUID="c4efa838-4f42-41e3-b12c-4574585bd8f4" containerName="route-controller-manager" containerID="cri-o://24055d9a4fe4ecdd8a9c53f05703e7a5f45b0adad3c69d4a6764d0d2323f230d" gracePeriod=30 Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.852902 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.895506 4808 patch_prober.go:28] interesting pod/route-controller-manager-688c64f9c6-dz2zb container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.49:8443/healthz\": read tcp 10.217.0.2:49070->10.217.0.49:8443: read: connection reset by peer" start-of-body= Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.895584 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" podUID="c4efa838-4f42-41e3-b12c-4574585bd8f4" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.49:8443/healthz\": read tcp 10.217.0.2:49070->10.217.0.49:8443: read: connection reset by peer" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.896373 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" podStartSLOduration=2.896359724 podStartE2EDuration="2.896359724s" podCreationTimestamp="2025-11-24 17:39:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:39:14.895821959 +0000 UTC m=+747.493489761" watchObservedRunningTime="2025-11-24 17:39:14.896359724 +0000 UTC m=+747.494027526" Nov 24 17:39:14 crc kubenswrapper[4808]: I1124 17:39:14.981185 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66bc449dc9-gnnjk"] Nov 24 17:39:15 crc kubenswrapper[4808]: I1124 17:39:15.866985 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" event={"ID":"36a6c0d1-772b-4ea4-a425-3a59b861f297","Type":"ContainerStarted","Data":"9356fddf1f9539b2b79515e513d48d09704222c6a6afb2180b66117eb1f4520c"} Nov 24 17:39:15 crc kubenswrapper[4808]: I1124 17:39:15.867703 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" event={"ID":"36a6c0d1-772b-4ea4-a425-3a59b861f297","Type":"ContainerStarted","Data":"6826fdc6e1136750cd394607a6f7c27925a653690afd7edd364e08ac2cc51e65"} Nov 24 17:39:15 crc kubenswrapper[4808]: I1124 17:39:15.867740 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:15 crc kubenswrapper[4808]: I1124 17:39:15.873122 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" Nov 24 17:39:15 crc kubenswrapper[4808]: I1124 17:39:15.876541 4808 generic.go:334] "Generic (PLEG): container finished" podID="c4efa838-4f42-41e3-b12c-4574585bd8f4" containerID="24055d9a4fe4ecdd8a9c53f05703e7a5f45b0adad3c69d4a6764d0d2323f230d" exitCode=0 Nov 24 17:39:15 crc kubenswrapper[4808]: I1124 17:39:15.876616 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" event={"ID":"c4efa838-4f42-41e3-b12c-4574585bd8f4","Type":"ContainerDied","Data":"24055d9a4fe4ecdd8a9c53f05703e7a5f45b0adad3c69d4a6764d0d2323f230d"} Nov 24 17:39:15 crc kubenswrapper[4808]: I1124 17:39:15.889074 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-66bc449dc9-gnnjk" podStartSLOduration=3.889052048 podStartE2EDuration="3.889052048s" podCreationTimestamp="2025-11-24 17:39:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:39:15.887567578 +0000 UTC m=+748.485235400" watchObservedRunningTime="2025-11-24 17:39:15.889052048 +0000 UTC m=+748.486719850" Nov 24 17:39:15 crc kubenswrapper[4808]: I1124 17:39:15.944239 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:15 crc kubenswrapper[4808]: I1124 17:39:15.976207 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k"] Nov 24 17:39:15 crc kubenswrapper[4808]: E1124 17:39:15.976482 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4efa838-4f42-41e3-b12c-4574585bd8f4" containerName="route-controller-manager" Nov 24 17:39:15 crc kubenswrapper[4808]: I1124 17:39:15.976496 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4efa838-4f42-41e3-b12c-4574585bd8f4" containerName="route-controller-manager" Nov 24 17:39:15 crc kubenswrapper[4808]: I1124 17:39:15.976634 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4efa838-4f42-41e3-b12c-4574585bd8f4" containerName="route-controller-manager" Nov 24 17:39:15 crc kubenswrapper[4808]: I1124 17:39:15.977095 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:15 crc kubenswrapper[4808]: I1124 17:39:15.990947 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k"] Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.084935 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4efa838-4f42-41e3-b12c-4574585bd8f4-client-ca\") pod \"c4efa838-4f42-41e3-b12c-4574585bd8f4\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.085083 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4efa838-4f42-41e3-b12c-4574585bd8f4-serving-cert\") pod \"c4efa838-4f42-41e3-b12c-4574585bd8f4\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.085137 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4efa838-4f42-41e3-b12c-4574585bd8f4-config\") pod \"c4efa838-4f42-41e3-b12c-4574585bd8f4\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.085166 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcm9m\" (UniqueName: \"kubernetes.io/projected/c4efa838-4f42-41e3-b12c-4574585bd8f4-kube-api-access-fcm9m\") pod \"c4efa838-4f42-41e3-b12c-4574585bd8f4\" (UID: \"c4efa838-4f42-41e3-b12c-4574585bd8f4\") " Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.085372 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtbqc\" (UniqueName: \"kubernetes.io/projected/c9999fb4-6965-4921-b43b-d202df093858-kube-api-access-qtbqc\") pod \"route-controller-manager-5698c99d8b-8lm5k\" (UID: \"c9999fb4-6965-4921-b43b-d202df093858\") " pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.085398 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9999fb4-6965-4921-b43b-d202df093858-config\") pod \"route-controller-manager-5698c99d8b-8lm5k\" (UID: \"c9999fb4-6965-4921-b43b-d202df093858\") " pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.085418 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9999fb4-6965-4921-b43b-d202df093858-serving-cert\") pod \"route-controller-manager-5698c99d8b-8lm5k\" (UID: \"c9999fb4-6965-4921-b43b-d202df093858\") " pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.086122 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4efa838-4f42-41e3-b12c-4574585bd8f4-client-ca" (OuterVolumeSpecName: "client-ca") pod "c4efa838-4f42-41e3-b12c-4574585bd8f4" (UID: "c4efa838-4f42-41e3-b12c-4574585bd8f4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.086198 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4efa838-4f42-41e3-b12c-4574585bd8f4-config" (OuterVolumeSpecName: "config") pod "c4efa838-4f42-41e3-b12c-4574585bd8f4" (UID: "c4efa838-4f42-41e3-b12c-4574585bd8f4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.086517 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c9999fb4-6965-4921-b43b-d202df093858-client-ca\") pod \"route-controller-manager-5698c99d8b-8lm5k\" (UID: \"c9999fb4-6965-4921-b43b-d202df093858\") " pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.086617 4808 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4efa838-4f42-41e3-b12c-4574585bd8f4-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.086639 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4efa838-4f42-41e3-b12c-4574585bd8f4-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.092412 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4efa838-4f42-41e3-b12c-4574585bd8f4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c4efa838-4f42-41e3-b12c-4574585bd8f4" (UID: "c4efa838-4f42-41e3-b12c-4574585bd8f4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.101940 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4efa838-4f42-41e3-b12c-4574585bd8f4-kube-api-access-fcm9m" (OuterVolumeSpecName: "kube-api-access-fcm9m") pod "c4efa838-4f42-41e3-b12c-4574585bd8f4" (UID: "c4efa838-4f42-41e3-b12c-4574585bd8f4"). InnerVolumeSpecName "kube-api-access-fcm9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.188625 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c9999fb4-6965-4921-b43b-d202df093858-client-ca\") pod \"route-controller-manager-5698c99d8b-8lm5k\" (UID: \"c9999fb4-6965-4921-b43b-d202df093858\") " pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.189252 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtbqc\" (UniqueName: \"kubernetes.io/projected/c9999fb4-6965-4921-b43b-d202df093858-kube-api-access-qtbqc\") pod \"route-controller-manager-5698c99d8b-8lm5k\" (UID: \"c9999fb4-6965-4921-b43b-d202df093858\") " pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.189284 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9999fb4-6965-4921-b43b-d202df093858-config\") pod \"route-controller-manager-5698c99d8b-8lm5k\" (UID: \"c9999fb4-6965-4921-b43b-d202df093858\") " pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.189309 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9999fb4-6965-4921-b43b-d202df093858-serving-cert\") pod \"route-controller-manager-5698c99d8b-8lm5k\" (UID: \"c9999fb4-6965-4921-b43b-d202df093858\") " pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.189722 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4efa838-4f42-41e3-b12c-4574585bd8f4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.189762 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcm9m\" (UniqueName: \"kubernetes.io/projected/c4efa838-4f42-41e3-b12c-4574585bd8f4-kube-api-access-fcm9m\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.189749 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c9999fb4-6965-4921-b43b-d202df093858-client-ca\") pod \"route-controller-manager-5698c99d8b-8lm5k\" (UID: \"c9999fb4-6965-4921-b43b-d202df093858\") " pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.190745 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9999fb4-6965-4921-b43b-d202df093858-config\") pod \"route-controller-manager-5698c99d8b-8lm5k\" (UID: \"c9999fb4-6965-4921-b43b-d202df093858\") " pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.193086 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9999fb4-6965-4921-b43b-d202df093858-serving-cert\") pod \"route-controller-manager-5698c99d8b-8lm5k\" (UID: \"c9999fb4-6965-4921-b43b-d202df093858\") " pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.213256 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtbqc\" (UniqueName: \"kubernetes.io/projected/c9999fb4-6965-4921-b43b-d202df093858-kube-api-access-qtbqc\") pod \"route-controller-manager-5698c99d8b-8lm5k\" (UID: \"c9999fb4-6965-4921-b43b-d202df093858\") " pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.293138 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.753786 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k"] Nov 24 17:39:16 crc kubenswrapper[4808]: W1124 17:39:16.764079 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9999fb4_6965_4921_b43b_d202df093858.slice/crio-2785f9376396e33a08a06bc508c1f626007310e67e38f97769321f06ff333fc1 WatchSource:0}: Error finding container 2785f9376396e33a08a06bc508c1f626007310e67e38f97769321f06ff333fc1: Status 404 returned error can't find the container with id 2785f9376396e33a08a06bc508c1f626007310e67e38f97769321f06ff333fc1 Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.885203 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.885236 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb" event={"ID":"c4efa838-4f42-41e3-b12c-4574585bd8f4","Type":"ContainerDied","Data":"a8bc4aa9005c29b6c79dfb2baed3d21938d47ddd52189ce025768db63962cf58"} Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.885767 4808 scope.go:117] "RemoveContainer" containerID="24055d9a4fe4ecdd8a9c53f05703e7a5f45b0adad3c69d4a6764d0d2323f230d" Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.886691 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" event={"ID":"c9999fb4-6965-4921-b43b-d202df093858","Type":"ContainerStarted","Data":"2785f9376396e33a08a06bc508c1f626007310e67e38f97769321f06ff333fc1"} Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.912902 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb"] Nov 24 17:39:16 crc kubenswrapper[4808]: I1124 17:39:16.919852 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-688c64f9c6-dz2zb"] Nov 24 17:39:17 crc kubenswrapper[4808]: I1124 17:39:17.569158 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-56c4766567-nztcn" Nov 24 17:39:17 crc kubenswrapper[4808]: I1124 17:39:17.902536 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" event={"ID":"c9999fb4-6965-4921-b43b-d202df093858","Type":"ContainerStarted","Data":"ab628a873ce22127fd15e1ebcb9795b4035da2b0d25618d94b516e766beed0b8"} Nov 24 17:39:17 crc kubenswrapper[4808]: I1124 17:39:17.905240 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:17 crc kubenswrapper[4808]: I1124 17:39:17.915046 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" Nov 24 17:39:17 crc kubenswrapper[4808]: I1124 17:39:17.922130 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5698c99d8b-8lm5k" podStartSLOduration=3.922107902 podStartE2EDuration="3.922107902s" podCreationTimestamp="2025-11-24 17:39:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:39:17.921039032 +0000 UTC m=+750.518706844" watchObservedRunningTime="2025-11-24 17:39:17.922107902 +0000 UTC m=+750.519775694" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.291601 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-b46nx"] Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.294854 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.300885 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.301462 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-dgtvq" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.303493 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.305203 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-9qp89"] Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.306500 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9qp89" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.308106 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.316243 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-9qp89"] Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.357561 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4efa838-4f42-41e3-b12c-4574585bd8f4" path="/var/lib/kubelet/pods/c4efa838-4f42-41e3-b12c-4574585bd8f4/volumes" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.402110 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-6mvpd"] Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.403273 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6mvpd" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.407336 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.407839 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-8sqvs" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.409721 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.411847 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.426194 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0e562199-e728-4698-8072-27b477593a5f-frr-conf\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.426265 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0e562199-e728-4698-8072-27b477593a5f-metrics\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.426427 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz9qk\" (UniqueName: \"kubernetes.io/projected/0e562199-e728-4698-8072-27b477593a5f-kube-api-access-gz9qk\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.426500 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0e562199-e728-4698-8072-27b477593a5f-reloader\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.426603 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0e562199-e728-4698-8072-27b477593a5f-metrics-certs\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.426633 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0e562199-e728-4698-8072-27b477593a5f-frr-startup\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.426710 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86gtb\" (UniqueName: \"kubernetes.io/projected/f5d3c911-b65c-46dc-a43a-60247b4d10f3-kube-api-access-86gtb\") pod \"frr-k8s-webhook-server-6998585d5-9qp89\" (UID: \"f5d3c911-b65c-46dc-a43a-60247b4d10f3\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9qp89" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.426803 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0e562199-e728-4698-8072-27b477593a5f-frr-sockets\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.426888 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5d3c911-b65c-46dc-a43a-60247b4d10f3-cert\") pod \"frr-k8s-webhook-server-6998585d5-9qp89\" (UID: \"f5d3c911-b65c-46dc-a43a-60247b4d10f3\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9qp89" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.440208 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-88twg"] Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.441687 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-88twg" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.444060 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.458186 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-88twg"] Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.527953 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0e562199-e728-4698-8072-27b477593a5f-frr-sockets\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.528039 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5d3c911-b65c-46dc-a43a-60247b4d10f3-cert\") pod \"frr-k8s-webhook-server-6998585d5-9qp89\" (UID: \"f5d3c911-b65c-46dc-a43a-60247b4d10f3\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9qp89" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.528089 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zdfw\" (UniqueName: \"kubernetes.io/projected/9afaca72-2a60-4b1f-a377-4510b24b887e-kube-api-access-6zdfw\") pod \"controller-6c7b4b5f48-88twg\" (UID: \"9afaca72-2a60-4b1f-a377-4510b24b887e\") " pod="metallb-system/controller-6c7b4b5f48-88twg" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.528164 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0e562199-e728-4698-8072-27b477593a5f-frr-conf\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.528198 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0e562199-e728-4698-8072-27b477593a5f-metrics\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: E1124 17:39:18.528271 4808 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.528316 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6f3367d5-54fd-4556-b1ba-5311e15433f3-metallb-excludel2\") pod \"speaker-6mvpd\" (UID: \"6f3367d5-54fd-4556-b1ba-5311e15433f3\") " pod="metallb-system/speaker-6mvpd" Nov 24 17:39:18 crc kubenswrapper[4808]: E1124 17:39:18.528358 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5d3c911-b65c-46dc-a43a-60247b4d10f3-cert podName:f5d3c911-b65c-46dc-a43a-60247b4d10f3 nodeName:}" failed. No retries permitted until 2025-11-24 17:39:19.02833074 +0000 UTC m=+751.625998542 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f5d3c911-b65c-46dc-a43a-60247b4d10f3-cert") pod "frr-k8s-webhook-server-6998585d5-9qp89" (UID: "f5d3c911-b65c-46dc-a43a-60247b4d10f3") : secret "frr-k8s-webhook-server-cert" not found Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.528385 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6f3367d5-54fd-4556-b1ba-5311e15433f3-metrics-certs\") pod \"speaker-6mvpd\" (UID: \"6f3367d5-54fd-4556-b1ba-5311e15433f3\") " pod="metallb-system/speaker-6mvpd" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.528419 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn445\" (UniqueName: \"kubernetes.io/projected/6f3367d5-54fd-4556-b1ba-5311e15433f3-kube-api-access-fn445\") pod \"speaker-6mvpd\" (UID: \"6f3367d5-54fd-4556-b1ba-5311e15433f3\") " pod="metallb-system/speaker-6mvpd" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.528514 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz9qk\" (UniqueName: \"kubernetes.io/projected/0e562199-e728-4698-8072-27b477593a5f-kube-api-access-gz9qk\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.528543 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0e562199-e728-4698-8072-27b477593a5f-reloader\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.528575 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0e562199-e728-4698-8072-27b477593a5f-metrics-certs\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.528661 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0e562199-e728-4698-8072-27b477593a5f-frr-sockets\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: E1124 17:39:18.528736 4808 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.528702 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6f3367d5-54fd-4556-b1ba-5311e15433f3-memberlist\") pod \"speaker-6mvpd\" (UID: \"6f3367d5-54fd-4556-b1ba-5311e15433f3\") " pod="metallb-system/speaker-6mvpd" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.528807 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0e562199-e728-4698-8072-27b477593a5f-frr-startup\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: E1124 17:39:18.528876 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0e562199-e728-4698-8072-27b477593a5f-metrics-certs podName:0e562199-e728-4698-8072-27b477593a5f nodeName:}" failed. No retries permitted until 2025-11-24 17:39:19.028829495 +0000 UTC m=+751.626497377 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0e562199-e728-4698-8072-27b477593a5f-metrics-certs") pod "frr-k8s-b46nx" (UID: "0e562199-e728-4698-8072-27b477593a5f") : secret "frr-k8s-certs-secret" not found Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.528878 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0e562199-e728-4698-8072-27b477593a5f-frr-conf\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.528938 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86gtb\" (UniqueName: \"kubernetes.io/projected/f5d3c911-b65c-46dc-a43a-60247b4d10f3-kube-api-access-86gtb\") pod \"frr-k8s-webhook-server-6998585d5-9qp89\" (UID: \"f5d3c911-b65c-46dc-a43a-60247b4d10f3\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9qp89" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.529000 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9afaca72-2a60-4b1f-a377-4510b24b887e-cert\") pod \"controller-6c7b4b5f48-88twg\" (UID: \"9afaca72-2a60-4b1f-a377-4510b24b887e\") " pod="metallb-system/controller-6c7b4b5f48-88twg" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.529052 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9afaca72-2a60-4b1f-a377-4510b24b887e-metrics-certs\") pod \"controller-6c7b4b5f48-88twg\" (UID: \"9afaca72-2a60-4b1f-a377-4510b24b887e\") " pod="metallb-system/controller-6c7b4b5f48-88twg" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.529055 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0e562199-e728-4698-8072-27b477593a5f-metrics\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.529812 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0e562199-e728-4698-8072-27b477593a5f-reloader\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.529999 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0e562199-e728-4698-8072-27b477593a5f-frr-startup\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.559281 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86gtb\" (UniqueName: \"kubernetes.io/projected/f5d3c911-b65c-46dc-a43a-60247b4d10f3-kube-api-access-86gtb\") pod \"frr-k8s-webhook-server-6998585d5-9qp89\" (UID: \"f5d3c911-b65c-46dc-a43a-60247b4d10f3\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9qp89" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.560563 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz9qk\" (UniqueName: \"kubernetes.io/projected/0e562199-e728-4698-8072-27b477593a5f-kube-api-access-gz9qk\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.630133 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6f3367d5-54fd-4556-b1ba-5311e15433f3-metallb-excludel2\") pod \"speaker-6mvpd\" (UID: \"6f3367d5-54fd-4556-b1ba-5311e15433f3\") " pod="metallb-system/speaker-6mvpd" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.630966 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6f3367d5-54fd-4556-b1ba-5311e15433f3-metrics-certs\") pod \"speaker-6mvpd\" (UID: \"6f3367d5-54fd-4556-b1ba-5311e15433f3\") " pod="metallb-system/speaker-6mvpd" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.630917 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6f3367d5-54fd-4556-b1ba-5311e15433f3-metallb-excludel2\") pod \"speaker-6mvpd\" (UID: \"6f3367d5-54fd-4556-b1ba-5311e15433f3\") " pod="metallb-system/speaker-6mvpd" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.631056 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn445\" (UniqueName: \"kubernetes.io/projected/6f3367d5-54fd-4556-b1ba-5311e15433f3-kube-api-access-fn445\") pod \"speaker-6mvpd\" (UID: \"6f3367d5-54fd-4556-b1ba-5311e15433f3\") " pod="metallb-system/speaker-6mvpd" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.631142 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6f3367d5-54fd-4556-b1ba-5311e15433f3-memberlist\") pod \"speaker-6mvpd\" (UID: \"6f3367d5-54fd-4556-b1ba-5311e15433f3\") " pod="metallb-system/speaker-6mvpd" Nov 24 17:39:18 crc kubenswrapper[4808]: E1124 17:39:18.631315 4808 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 17:39:18 crc kubenswrapper[4808]: E1124 17:39:18.631428 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f3367d5-54fd-4556-b1ba-5311e15433f3-memberlist podName:6f3367d5-54fd-4556-b1ba-5311e15433f3 nodeName:}" failed. No retries permitted until 2025-11-24 17:39:19.131409927 +0000 UTC m=+751.729077729 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/6f3367d5-54fd-4556-b1ba-5311e15433f3-memberlist") pod "speaker-6mvpd" (UID: "6f3367d5-54fd-4556-b1ba-5311e15433f3") : secret "metallb-memberlist" not found Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.631906 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9afaca72-2a60-4b1f-a377-4510b24b887e-cert\") pod \"controller-6c7b4b5f48-88twg\" (UID: \"9afaca72-2a60-4b1f-a377-4510b24b887e\") " pod="metallb-system/controller-6c7b4b5f48-88twg" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.632538 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9afaca72-2a60-4b1f-a377-4510b24b887e-metrics-certs\") pod \"controller-6c7b4b5f48-88twg\" (UID: \"9afaca72-2a60-4b1f-a377-4510b24b887e\") " pod="metallb-system/controller-6c7b4b5f48-88twg" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.632620 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zdfw\" (UniqueName: \"kubernetes.io/projected/9afaca72-2a60-4b1f-a377-4510b24b887e-kube-api-access-6zdfw\") pod \"controller-6c7b4b5f48-88twg\" (UID: \"9afaca72-2a60-4b1f-a377-4510b24b887e\") " pod="metallb-system/controller-6c7b4b5f48-88twg" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.636379 4808 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.636943 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6f3367d5-54fd-4556-b1ba-5311e15433f3-metrics-certs\") pod \"speaker-6mvpd\" (UID: \"6f3367d5-54fd-4556-b1ba-5311e15433f3\") " pod="metallb-system/speaker-6mvpd" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.638793 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9afaca72-2a60-4b1f-a377-4510b24b887e-metrics-certs\") pod \"controller-6c7b4b5f48-88twg\" (UID: \"9afaca72-2a60-4b1f-a377-4510b24b887e\") " pod="metallb-system/controller-6c7b4b5f48-88twg" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.646485 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9afaca72-2a60-4b1f-a377-4510b24b887e-cert\") pod \"controller-6c7b4b5f48-88twg\" (UID: \"9afaca72-2a60-4b1f-a377-4510b24b887e\") " pod="metallb-system/controller-6c7b4b5f48-88twg" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.653624 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zdfw\" (UniqueName: \"kubernetes.io/projected/9afaca72-2a60-4b1f-a377-4510b24b887e-kube-api-access-6zdfw\") pod \"controller-6c7b4b5f48-88twg\" (UID: \"9afaca72-2a60-4b1f-a377-4510b24b887e\") " pod="metallb-system/controller-6c7b4b5f48-88twg" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.654241 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn445\" (UniqueName: \"kubernetes.io/projected/6f3367d5-54fd-4556-b1ba-5311e15433f3-kube-api-access-fn445\") pod \"speaker-6mvpd\" (UID: \"6f3367d5-54fd-4556-b1ba-5311e15433f3\") " pod="metallb-system/speaker-6mvpd" Nov 24 17:39:18 crc kubenswrapper[4808]: I1124 17:39:18.763627 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-88twg" Nov 24 17:39:19 crc kubenswrapper[4808]: I1124 17:39:19.039810 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5d3c911-b65c-46dc-a43a-60247b4d10f3-cert\") pod \"frr-k8s-webhook-server-6998585d5-9qp89\" (UID: \"f5d3c911-b65c-46dc-a43a-60247b4d10f3\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9qp89" Nov 24 17:39:19 crc kubenswrapper[4808]: I1124 17:39:19.039974 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0e562199-e728-4698-8072-27b477593a5f-metrics-certs\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:19 crc kubenswrapper[4808]: I1124 17:39:19.046753 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0e562199-e728-4698-8072-27b477593a5f-metrics-certs\") pod \"frr-k8s-b46nx\" (UID: \"0e562199-e728-4698-8072-27b477593a5f\") " pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:19 crc kubenswrapper[4808]: I1124 17:39:19.046908 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5d3c911-b65c-46dc-a43a-60247b4d10f3-cert\") pod \"frr-k8s-webhook-server-6998585d5-9qp89\" (UID: \"f5d3c911-b65c-46dc-a43a-60247b4d10f3\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9qp89" Nov 24 17:39:19 crc kubenswrapper[4808]: I1124 17:39:19.141592 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6f3367d5-54fd-4556-b1ba-5311e15433f3-memberlist\") pod \"speaker-6mvpd\" (UID: \"6f3367d5-54fd-4556-b1ba-5311e15433f3\") " pod="metallb-system/speaker-6mvpd" Nov 24 17:39:19 crc kubenswrapper[4808]: E1124 17:39:19.141837 4808 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 17:39:19 crc kubenswrapper[4808]: E1124 17:39:19.142381 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f3367d5-54fd-4556-b1ba-5311e15433f3-memberlist podName:6f3367d5-54fd-4556-b1ba-5311e15433f3 nodeName:}" failed. No retries permitted until 2025-11-24 17:39:20.14235137 +0000 UTC m=+752.740019172 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/6f3367d5-54fd-4556-b1ba-5311e15433f3-memberlist") pod "speaker-6mvpd" (UID: "6f3367d5-54fd-4556-b1ba-5311e15433f3") : secret "metallb-memberlist" not found Nov 24 17:39:19 crc kubenswrapper[4808]: I1124 17:39:19.222131 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:19 crc kubenswrapper[4808]: I1124 17:39:19.229039 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9qp89" Nov 24 17:39:19 crc kubenswrapper[4808]: I1124 17:39:19.346003 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-88twg"] Nov 24 17:39:19 crc kubenswrapper[4808]: I1124 17:39:19.699567 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-9qp89"] Nov 24 17:39:19 crc kubenswrapper[4808]: W1124 17:39:19.705338 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5d3c911_b65c_46dc_a43a_60247b4d10f3.slice/crio-18d35b903cfdfc19775a860d27a019a4891bff73472b45a85d9346554ff6419b WatchSource:0}: Error finding container 18d35b903cfdfc19775a860d27a019a4891bff73472b45a85d9346554ff6419b: Status 404 returned error can't find the container with id 18d35b903cfdfc19775a860d27a019a4891bff73472b45a85d9346554ff6419b Nov 24 17:39:19 crc kubenswrapper[4808]: I1124 17:39:19.915320 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-88twg" event={"ID":"9afaca72-2a60-4b1f-a377-4510b24b887e","Type":"ContainerStarted","Data":"6cc033944fd5874c49cfbbb32fc513c22ebfdb3ae4ca3ab2a12aae8f21c50c91"} Nov 24 17:39:19 crc kubenswrapper[4808]: I1124 17:39:19.915387 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-88twg" event={"ID":"9afaca72-2a60-4b1f-a377-4510b24b887e","Type":"ContainerStarted","Data":"1f8c7fd2fd1170eb803a6b8adbee1188ba07a9e95b61c9c92dd3b978b820fac1"} Nov 24 17:39:19 crc kubenswrapper[4808]: I1124 17:39:19.916322 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9qp89" event={"ID":"f5d3c911-b65c-46dc-a43a-60247b4d10f3","Type":"ContainerStarted","Data":"18d35b903cfdfc19775a860d27a019a4891bff73472b45a85d9346554ff6419b"} Nov 24 17:39:19 crc kubenswrapper[4808]: I1124 17:39:19.918420 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b46nx" event={"ID":"0e562199-e728-4698-8072-27b477593a5f","Type":"ContainerStarted","Data":"0a456cdee4f3437d8a30011016ad0ce99185ea39c5aeab83b5d9cb8ea7fb8a6f"} Nov 24 17:39:20 crc kubenswrapper[4808]: I1124 17:39:20.162450 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6f3367d5-54fd-4556-b1ba-5311e15433f3-memberlist\") pod \"speaker-6mvpd\" (UID: \"6f3367d5-54fd-4556-b1ba-5311e15433f3\") " pod="metallb-system/speaker-6mvpd" Nov 24 17:39:20 crc kubenswrapper[4808]: I1124 17:39:20.171148 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6f3367d5-54fd-4556-b1ba-5311e15433f3-memberlist\") pod \"speaker-6mvpd\" (UID: \"6f3367d5-54fd-4556-b1ba-5311e15433f3\") " pod="metallb-system/speaker-6mvpd" Nov 24 17:39:20 crc kubenswrapper[4808]: I1124 17:39:20.220196 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6mvpd" Nov 24 17:39:20 crc kubenswrapper[4808]: W1124 17:39:20.245184 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f3367d5_54fd_4556_b1ba_5311e15433f3.slice/crio-c3b2a59f5650d16f1869403717471f9ab7760b24638f2f83e2e797d52018b574 WatchSource:0}: Error finding container c3b2a59f5650d16f1869403717471f9ab7760b24638f2f83e2e797d52018b574: Status 404 returned error can't find the container with id c3b2a59f5650d16f1869403717471f9ab7760b24638f2f83e2e797d52018b574 Nov 24 17:39:20 crc kubenswrapper[4808]: I1124 17:39:20.931393 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-88twg" event={"ID":"9afaca72-2a60-4b1f-a377-4510b24b887e","Type":"ContainerStarted","Data":"a29d92b47f17f8f15498c8ca1adced66f82a6ed9b160b9508ebb338f851a27f2"} Nov 24 17:39:20 crc kubenswrapper[4808]: I1124 17:39:20.931805 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-88twg" Nov 24 17:39:20 crc kubenswrapper[4808]: I1124 17:39:20.936737 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6mvpd" event={"ID":"6f3367d5-54fd-4556-b1ba-5311e15433f3","Type":"ContainerStarted","Data":"1bc38e9d588cf49f92cc7dd49f6d21ad4c80b87cfd590b112c2bf1163bdf6e6f"} Nov 24 17:39:20 crc kubenswrapper[4808]: I1124 17:39:20.936793 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6mvpd" event={"ID":"6f3367d5-54fd-4556-b1ba-5311e15433f3","Type":"ContainerStarted","Data":"c3b2a59f5650d16f1869403717471f9ab7760b24638f2f83e2e797d52018b574"} Nov 24 17:39:20 crc kubenswrapper[4808]: I1124 17:39:20.980486 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-88twg" podStartSLOduration=2.98045573 podStartE2EDuration="2.98045573s" podCreationTimestamp="2025-11-24 17:39:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:39:20.978361648 +0000 UTC m=+753.576029460" watchObservedRunningTime="2025-11-24 17:39:20.98045573 +0000 UTC m=+753.578123532" Nov 24 17:39:21 crc kubenswrapper[4808]: I1124 17:39:21.959249 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6mvpd" event={"ID":"6f3367d5-54fd-4556-b1ba-5311e15433f3","Type":"ContainerStarted","Data":"4016a398a193baf04e2606eec0c0585d45417ba56a5d105c7f54f19740082899"} Nov 24 17:39:21 crc kubenswrapper[4808]: I1124 17:39:21.959631 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-6mvpd" Nov 24 17:39:21 crc kubenswrapper[4808]: I1124 17:39:21.983083 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-6mvpd" podStartSLOduration=3.983056337 podStartE2EDuration="3.983056337s" podCreationTimestamp="2025-11-24 17:39:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:39:21.980973266 +0000 UTC m=+754.578641068" watchObservedRunningTime="2025-11-24 17:39:21.983056337 +0000 UTC m=+754.580724139" Nov 24 17:39:24 crc kubenswrapper[4808]: I1124 17:39:24.819104 4808 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 17:39:30 crc kubenswrapper[4808]: I1124 17:39:30.011476 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9qp89" event={"ID":"f5d3c911-b65c-46dc-a43a-60247b4d10f3","Type":"ContainerStarted","Data":"ef8b03791713ce0859ae882e27fcf3d28ed847417dc2d802ce4fddea546c36fa"} Nov 24 17:39:30 crc kubenswrapper[4808]: I1124 17:39:30.012152 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9qp89" Nov 24 17:39:30 crc kubenswrapper[4808]: I1124 17:39:30.013234 4808 generic.go:334] "Generic (PLEG): container finished" podID="0e562199-e728-4698-8072-27b477593a5f" containerID="cacf454ca17d4b521afb837a3ae7391ca4411753bd22bca7dac8c5fa716430e2" exitCode=0 Nov 24 17:39:30 crc kubenswrapper[4808]: I1124 17:39:30.013256 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b46nx" event={"ID":"0e562199-e728-4698-8072-27b477593a5f","Type":"ContainerDied","Data":"cacf454ca17d4b521afb837a3ae7391ca4411753bd22bca7dac8c5fa716430e2"} Nov 24 17:39:30 crc kubenswrapper[4808]: I1124 17:39:30.033662 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9qp89" podStartSLOduration=2.053405189 podStartE2EDuration="12.033645479s" podCreationTimestamp="2025-11-24 17:39:18 +0000 UTC" firstStartedPulling="2025-11-24 17:39:19.708056515 +0000 UTC m=+752.305724317" lastFinishedPulling="2025-11-24 17:39:29.688296805 +0000 UTC m=+762.285964607" observedRunningTime="2025-11-24 17:39:30.02995654 +0000 UTC m=+762.627624342" watchObservedRunningTime="2025-11-24 17:39:30.033645479 +0000 UTC m=+762.631313281" Nov 24 17:39:30 crc kubenswrapper[4808]: I1124 17:39:30.225724 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-6mvpd" Nov 24 17:39:31 crc kubenswrapper[4808]: I1124 17:39:31.021097 4808 generic.go:334] "Generic (PLEG): container finished" podID="0e562199-e728-4698-8072-27b477593a5f" containerID="711a618130b16f09847249c595b599890700d0cc8cde878549654d4032a5b959" exitCode=0 Nov 24 17:39:31 crc kubenswrapper[4808]: I1124 17:39:31.021195 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b46nx" event={"ID":"0e562199-e728-4698-8072-27b477593a5f","Type":"ContainerDied","Data":"711a618130b16f09847249c595b599890700d0cc8cde878549654d4032a5b959"} Nov 24 17:39:32 crc kubenswrapper[4808]: I1124 17:39:32.032036 4808 generic.go:334] "Generic (PLEG): container finished" podID="0e562199-e728-4698-8072-27b477593a5f" containerID="3f388ab9c62902bfc22d6b17adbd1bca3b029f0fb0090894e490dcc8436448f9" exitCode=0 Nov 24 17:39:32 crc kubenswrapper[4808]: I1124 17:39:32.032066 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b46nx" event={"ID":"0e562199-e728-4698-8072-27b477593a5f","Type":"ContainerDied","Data":"3f388ab9c62902bfc22d6b17adbd1bca3b029f0fb0090894e490dcc8436448f9"} Nov 24 17:39:33 crc kubenswrapper[4808]: I1124 17:39:33.043254 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b46nx" event={"ID":"0e562199-e728-4698-8072-27b477593a5f","Type":"ContainerStarted","Data":"611cb8adc5f00a4a3108efc9288434e575ec1a3b27015c8a1fc4e9aa9ab2cca7"} Nov 24 17:39:33 crc kubenswrapper[4808]: I1124 17:39:33.043821 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b46nx" event={"ID":"0e562199-e728-4698-8072-27b477593a5f","Type":"ContainerStarted","Data":"41f3c15e2647df14d90ccf1bb24eb40eb63069a2e5fd687c11e5bc1d7a43a83a"} Nov 24 17:39:33 crc kubenswrapper[4808]: I1124 17:39:33.043838 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b46nx" event={"ID":"0e562199-e728-4698-8072-27b477593a5f","Type":"ContainerStarted","Data":"591f97b50bf12d7898c8dece4589cea4dc7db1afb7b36e7a6daedb36a66e0f4c"} Nov 24 17:39:33 crc kubenswrapper[4808]: I1124 17:39:33.316909 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-kn28w"] Nov 24 17:39:33 crc kubenswrapper[4808]: I1124 17:39:33.317701 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kn28w" Nov 24 17:39:33 crc kubenswrapper[4808]: I1124 17:39:33.319812 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 24 17:39:33 crc kubenswrapper[4808]: I1124 17:39:33.321747 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 24 17:39:33 crc kubenswrapper[4808]: I1124 17:39:33.329306 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kn28w"] Nov 24 17:39:33 crc kubenswrapper[4808]: I1124 17:39:33.400413 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rxrt\" (UniqueName: \"kubernetes.io/projected/666531cc-9f5f-4f42-8c28-e3de1f361257-kube-api-access-2rxrt\") pod \"openstack-operator-index-kn28w\" (UID: \"666531cc-9f5f-4f42-8c28-e3de1f361257\") " pod="openstack-operators/openstack-operator-index-kn28w" Nov 24 17:39:33 crc kubenswrapper[4808]: I1124 17:39:33.501737 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rxrt\" (UniqueName: \"kubernetes.io/projected/666531cc-9f5f-4f42-8c28-e3de1f361257-kube-api-access-2rxrt\") pod \"openstack-operator-index-kn28w\" (UID: \"666531cc-9f5f-4f42-8c28-e3de1f361257\") " pod="openstack-operators/openstack-operator-index-kn28w" Nov 24 17:39:33 crc kubenswrapper[4808]: I1124 17:39:33.527182 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rxrt\" (UniqueName: \"kubernetes.io/projected/666531cc-9f5f-4f42-8c28-e3de1f361257-kube-api-access-2rxrt\") pod \"openstack-operator-index-kn28w\" (UID: \"666531cc-9f5f-4f42-8c28-e3de1f361257\") " pod="openstack-operators/openstack-operator-index-kn28w" Nov 24 17:39:33 crc kubenswrapper[4808]: I1124 17:39:33.635442 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kn28w" Nov 24 17:39:34 crc kubenswrapper[4808]: I1124 17:39:34.055004 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b46nx" event={"ID":"0e562199-e728-4698-8072-27b477593a5f","Type":"ContainerStarted","Data":"0a217a5dded2842d0e35e1db6d0302755568318a2657e81025eea424fa2197ba"} Nov 24 17:39:34 crc kubenswrapper[4808]: I1124 17:39:34.064617 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kn28w"] Nov 24 17:39:34 crc kubenswrapper[4808]: W1124 17:39:34.070181 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod666531cc_9f5f_4f42_8c28_e3de1f361257.slice/crio-e28d61c2cc894fcc6f3077300e8c878cf8be02f7da9833dbdba9b28f51ef2199 WatchSource:0}: Error finding container e28d61c2cc894fcc6f3077300e8c878cf8be02f7da9833dbdba9b28f51ef2199: Status 404 returned error can't find the container with id e28d61c2cc894fcc6f3077300e8c878cf8be02f7da9833dbdba9b28f51ef2199 Nov 24 17:39:35 crc kubenswrapper[4808]: I1124 17:39:35.065925 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kn28w" event={"ID":"666531cc-9f5f-4f42-8c28-e3de1f361257","Type":"ContainerStarted","Data":"e28d61c2cc894fcc6f3077300e8c878cf8be02f7da9833dbdba9b28f51ef2199"} Nov 24 17:39:35 crc kubenswrapper[4808]: I1124 17:39:35.072260 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b46nx" event={"ID":"0e562199-e728-4698-8072-27b477593a5f","Type":"ContainerStarted","Data":"9080be54022d24433109b904c7fe254430fb678746bc03cd592a392f3457cc9e"} Nov 24 17:39:35 crc kubenswrapper[4808]: I1124 17:39:35.072293 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b46nx" event={"ID":"0e562199-e728-4698-8072-27b477593a5f","Type":"ContainerStarted","Data":"4a0ab0e2371a17a8bd77c2f6a20d7af2fa90c8439bf729d0df277f62cd5702aa"} Nov 24 17:39:35 crc kubenswrapper[4808]: I1124 17:39:35.115919 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-b46nx" podStartSLOduration=7.065525949 podStartE2EDuration="17.115887113s" podCreationTimestamp="2025-11-24 17:39:18 +0000 UTC" firstStartedPulling="2025-11-24 17:39:19.661864685 +0000 UTC m=+752.259532487" lastFinishedPulling="2025-11-24 17:39:29.712225839 +0000 UTC m=+762.309893651" observedRunningTime="2025-11-24 17:39:35.110132303 +0000 UTC m=+767.707800115" watchObservedRunningTime="2025-11-24 17:39:35.115887113 +0000 UTC m=+767.713554915" Nov 24 17:39:36 crc kubenswrapper[4808]: I1124 17:39:36.080523 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:36 crc kubenswrapper[4808]: I1124 17:39:36.522947 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:39:36 crc kubenswrapper[4808]: I1124 17:39:36.523165 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:39:37 crc kubenswrapper[4808]: I1124 17:39:37.095431 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-kn28w"] Nov 24 17:39:37 crc kubenswrapper[4808]: I1124 17:39:37.706824 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-d6xmw"] Nov 24 17:39:37 crc kubenswrapper[4808]: I1124 17:39:37.708896 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-d6xmw" Nov 24 17:39:37 crc kubenswrapper[4808]: I1124 17:39:37.712895 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-d6xmw"] Nov 24 17:39:37 crc kubenswrapper[4808]: I1124 17:39:37.716411 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-nkrz2" Nov 24 17:39:37 crc kubenswrapper[4808]: I1124 17:39:37.873565 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsgj9\" (UniqueName: \"kubernetes.io/projected/24eb3123-6595-48d6-8d4a-f984d9a9e601-kube-api-access-gsgj9\") pod \"openstack-operator-index-d6xmw\" (UID: \"24eb3123-6595-48d6-8d4a-f984d9a9e601\") " pod="openstack-operators/openstack-operator-index-d6xmw" Nov 24 17:39:37 crc kubenswrapper[4808]: I1124 17:39:37.975355 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsgj9\" (UniqueName: \"kubernetes.io/projected/24eb3123-6595-48d6-8d4a-f984d9a9e601-kube-api-access-gsgj9\") pod \"openstack-operator-index-d6xmw\" (UID: \"24eb3123-6595-48d6-8d4a-f984d9a9e601\") " pod="openstack-operators/openstack-operator-index-d6xmw" Nov 24 17:39:37 crc kubenswrapper[4808]: I1124 17:39:37.998827 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsgj9\" (UniqueName: \"kubernetes.io/projected/24eb3123-6595-48d6-8d4a-f984d9a9e601-kube-api-access-gsgj9\") pod \"openstack-operator-index-d6xmw\" (UID: \"24eb3123-6595-48d6-8d4a-f984d9a9e601\") " pod="openstack-operators/openstack-operator-index-d6xmw" Nov 24 17:39:38 crc kubenswrapper[4808]: I1124 17:39:38.037108 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-d6xmw" Nov 24 17:39:38 crc kubenswrapper[4808]: I1124 17:39:38.475615 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-d6xmw"] Nov 24 17:39:38 crc kubenswrapper[4808]: I1124 17:39:38.769188 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-88twg" Nov 24 17:39:39 crc kubenswrapper[4808]: W1124 17:39:39.088771 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24eb3123_6595_48d6_8d4a_f984d9a9e601.slice/crio-483158af0214c38c4e3890bf2a837c56599fda826df6b93675586eefc3966042 WatchSource:0}: Error finding container 483158af0214c38c4e3890bf2a837c56599fda826df6b93675586eefc3966042: Status 404 returned error can't find the container with id 483158af0214c38c4e3890bf2a837c56599fda826df6b93675586eefc3966042 Nov 24 17:39:39 crc kubenswrapper[4808]: I1124 17:39:39.107621 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-d6xmw" event={"ID":"24eb3123-6595-48d6-8d4a-f984d9a9e601","Type":"ContainerStarted","Data":"483158af0214c38c4e3890bf2a837c56599fda826df6b93675586eefc3966042"} Nov 24 17:39:39 crc kubenswrapper[4808]: I1124 17:39:39.223146 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:39 crc kubenswrapper[4808]: I1124 17:39:39.237575 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9qp89" Nov 24 17:39:39 crc kubenswrapper[4808]: I1124 17:39:39.279497 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:41 crc kubenswrapper[4808]: I1124 17:39:41.505966 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9wdtd"] Nov 24 17:39:41 crc kubenswrapper[4808]: I1124 17:39:41.508012 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:41 crc kubenswrapper[4808]: I1124 17:39:41.517160 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9wdtd"] Nov 24 17:39:41 crc kubenswrapper[4808]: I1124 17:39:41.633512 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbb59\" (UniqueName: \"kubernetes.io/projected/74d302c4-75f5-4cd4-a72e-4c76961320c7-kube-api-access-sbb59\") pod \"community-operators-9wdtd\" (UID: \"74d302c4-75f5-4cd4-a72e-4c76961320c7\") " pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:41 crc kubenswrapper[4808]: I1124 17:39:41.633588 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74d302c4-75f5-4cd4-a72e-4c76961320c7-catalog-content\") pod \"community-operators-9wdtd\" (UID: \"74d302c4-75f5-4cd4-a72e-4c76961320c7\") " pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:41 crc kubenswrapper[4808]: I1124 17:39:41.633701 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74d302c4-75f5-4cd4-a72e-4c76961320c7-utilities\") pod \"community-operators-9wdtd\" (UID: \"74d302c4-75f5-4cd4-a72e-4c76961320c7\") " pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:41 crc kubenswrapper[4808]: I1124 17:39:41.735871 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74d302c4-75f5-4cd4-a72e-4c76961320c7-utilities\") pod \"community-operators-9wdtd\" (UID: \"74d302c4-75f5-4cd4-a72e-4c76961320c7\") " pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:41 crc kubenswrapper[4808]: I1124 17:39:41.735991 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbb59\" (UniqueName: \"kubernetes.io/projected/74d302c4-75f5-4cd4-a72e-4c76961320c7-kube-api-access-sbb59\") pod \"community-operators-9wdtd\" (UID: \"74d302c4-75f5-4cd4-a72e-4c76961320c7\") " pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:41 crc kubenswrapper[4808]: I1124 17:39:41.736050 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74d302c4-75f5-4cd4-a72e-4c76961320c7-catalog-content\") pod \"community-operators-9wdtd\" (UID: \"74d302c4-75f5-4cd4-a72e-4c76961320c7\") " pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:41 crc kubenswrapper[4808]: I1124 17:39:41.736509 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74d302c4-75f5-4cd4-a72e-4c76961320c7-utilities\") pod \"community-operators-9wdtd\" (UID: \"74d302c4-75f5-4cd4-a72e-4c76961320c7\") " pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:41 crc kubenswrapper[4808]: I1124 17:39:41.736603 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74d302c4-75f5-4cd4-a72e-4c76961320c7-catalog-content\") pod \"community-operators-9wdtd\" (UID: \"74d302c4-75f5-4cd4-a72e-4c76961320c7\") " pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:41 crc kubenswrapper[4808]: I1124 17:39:41.759152 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbb59\" (UniqueName: \"kubernetes.io/projected/74d302c4-75f5-4cd4-a72e-4c76961320c7-kube-api-access-sbb59\") pod \"community-operators-9wdtd\" (UID: \"74d302c4-75f5-4cd4-a72e-4c76961320c7\") " pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:41 crc kubenswrapper[4808]: I1124 17:39:41.844692 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:42 crc kubenswrapper[4808]: I1124 17:39:42.145211 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-d6xmw" event={"ID":"24eb3123-6595-48d6-8d4a-f984d9a9e601","Type":"ContainerStarted","Data":"b3f3f2b9d68684a066a43edb97284cae329d3c004eaead4e85798acbf812ba66"} Nov 24 17:39:42 crc kubenswrapper[4808]: I1124 17:39:42.147459 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kn28w" event={"ID":"666531cc-9f5f-4f42-8c28-e3de1f361257","Type":"ContainerStarted","Data":"657d77e2b9e4c869a1bf54ad0d20982235c5f5d2c5fb00eb31e1263727092cd1"} Nov 24 17:39:42 crc kubenswrapper[4808]: I1124 17:39:42.147695 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-kn28w" podUID="666531cc-9f5f-4f42-8c28-e3de1f361257" containerName="registry-server" containerID="cri-o://657d77e2b9e4c869a1bf54ad0d20982235c5f5d2c5fb00eb31e1263727092cd1" gracePeriod=2 Nov 24 17:39:42 crc kubenswrapper[4808]: I1124 17:39:42.170102 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-d6xmw" podStartSLOduration=2.622378724 podStartE2EDuration="5.17008023s" podCreationTimestamp="2025-11-24 17:39:37 +0000 UTC" firstStartedPulling="2025-11-24 17:39:39.091512135 +0000 UTC m=+771.689179937" lastFinishedPulling="2025-11-24 17:39:41.639213621 +0000 UTC m=+774.236881443" observedRunningTime="2025-11-24 17:39:42.167378451 +0000 UTC m=+774.765046253" watchObservedRunningTime="2025-11-24 17:39:42.17008023 +0000 UTC m=+774.767748032" Nov 24 17:39:42 crc kubenswrapper[4808]: I1124 17:39:42.380521 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-kn28w" podStartSLOduration=1.8156818289999999 podStartE2EDuration="9.380499839s" podCreationTimestamp="2025-11-24 17:39:33 +0000 UTC" firstStartedPulling="2025-11-24 17:39:34.073394681 +0000 UTC m=+766.671062483" lastFinishedPulling="2025-11-24 17:39:41.638212671 +0000 UTC m=+774.235880493" observedRunningTime="2025-11-24 17:39:42.18975525 +0000 UTC m=+774.787423052" watchObservedRunningTime="2025-11-24 17:39:42.380499839 +0000 UTC m=+774.978167641" Nov 24 17:39:42 crc kubenswrapper[4808]: I1124 17:39:42.381930 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9wdtd"] Nov 24 17:39:42 crc kubenswrapper[4808]: I1124 17:39:42.554405 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kn28w" Nov 24 17:39:42 crc kubenswrapper[4808]: I1124 17:39:42.657190 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rxrt\" (UniqueName: \"kubernetes.io/projected/666531cc-9f5f-4f42-8c28-e3de1f361257-kube-api-access-2rxrt\") pod \"666531cc-9f5f-4f42-8c28-e3de1f361257\" (UID: \"666531cc-9f5f-4f42-8c28-e3de1f361257\") " Nov 24 17:39:42 crc kubenswrapper[4808]: I1124 17:39:42.662797 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/666531cc-9f5f-4f42-8c28-e3de1f361257-kube-api-access-2rxrt" (OuterVolumeSpecName: "kube-api-access-2rxrt") pod "666531cc-9f5f-4f42-8c28-e3de1f361257" (UID: "666531cc-9f5f-4f42-8c28-e3de1f361257"). InnerVolumeSpecName "kube-api-access-2rxrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:39:42 crc kubenswrapper[4808]: I1124 17:39:42.759006 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rxrt\" (UniqueName: \"kubernetes.io/projected/666531cc-9f5f-4f42-8c28-e3de1f361257-kube-api-access-2rxrt\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:43 crc kubenswrapper[4808]: I1124 17:39:43.155759 4808 generic.go:334] "Generic (PLEG): container finished" podID="74d302c4-75f5-4cd4-a72e-4c76961320c7" containerID="9dda9bf93dd2c6f27fe90c8907e6ac375539628e44e075d7de3f5f9913bdea00" exitCode=0 Nov 24 17:39:43 crc kubenswrapper[4808]: I1124 17:39:43.155843 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wdtd" event={"ID":"74d302c4-75f5-4cd4-a72e-4c76961320c7","Type":"ContainerDied","Data":"9dda9bf93dd2c6f27fe90c8907e6ac375539628e44e075d7de3f5f9913bdea00"} Nov 24 17:39:43 crc kubenswrapper[4808]: I1124 17:39:43.156209 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wdtd" event={"ID":"74d302c4-75f5-4cd4-a72e-4c76961320c7","Type":"ContainerStarted","Data":"8d44aef9ee6ac165b931cae7aa0bba5184c759b39789cd8bee9dc36cb3087c18"} Nov 24 17:39:43 crc kubenswrapper[4808]: I1124 17:39:43.159101 4808 generic.go:334] "Generic (PLEG): container finished" podID="666531cc-9f5f-4f42-8c28-e3de1f361257" containerID="657d77e2b9e4c869a1bf54ad0d20982235c5f5d2c5fb00eb31e1263727092cd1" exitCode=0 Nov 24 17:39:43 crc kubenswrapper[4808]: I1124 17:39:43.159137 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kn28w" event={"ID":"666531cc-9f5f-4f42-8c28-e3de1f361257","Type":"ContainerDied","Data":"657d77e2b9e4c869a1bf54ad0d20982235c5f5d2c5fb00eb31e1263727092cd1"} Nov 24 17:39:43 crc kubenswrapper[4808]: I1124 17:39:43.159193 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kn28w" Nov 24 17:39:43 crc kubenswrapper[4808]: I1124 17:39:43.159223 4808 scope.go:117] "RemoveContainer" containerID="657d77e2b9e4c869a1bf54ad0d20982235c5f5d2c5fb00eb31e1263727092cd1" Nov 24 17:39:43 crc kubenswrapper[4808]: I1124 17:39:43.159201 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kn28w" event={"ID":"666531cc-9f5f-4f42-8c28-e3de1f361257","Type":"ContainerDied","Data":"e28d61c2cc894fcc6f3077300e8c878cf8be02f7da9833dbdba9b28f51ef2199"} Nov 24 17:39:43 crc kubenswrapper[4808]: I1124 17:39:43.179801 4808 scope.go:117] "RemoveContainer" containerID="657d77e2b9e4c869a1bf54ad0d20982235c5f5d2c5fb00eb31e1263727092cd1" Nov 24 17:39:43 crc kubenswrapper[4808]: E1124 17:39:43.180587 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"657d77e2b9e4c869a1bf54ad0d20982235c5f5d2c5fb00eb31e1263727092cd1\": container with ID starting with 657d77e2b9e4c869a1bf54ad0d20982235c5f5d2c5fb00eb31e1263727092cd1 not found: ID does not exist" containerID="657d77e2b9e4c869a1bf54ad0d20982235c5f5d2c5fb00eb31e1263727092cd1" Nov 24 17:39:43 crc kubenswrapper[4808]: I1124 17:39:43.180639 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"657d77e2b9e4c869a1bf54ad0d20982235c5f5d2c5fb00eb31e1263727092cd1"} err="failed to get container status \"657d77e2b9e4c869a1bf54ad0d20982235c5f5d2c5fb00eb31e1263727092cd1\": rpc error: code = NotFound desc = could not find container \"657d77e2b9e4c869a1bf54ad0d20982235c5f5d2c5fb00eb31e1263727092cd1\": container with ID starting with 657d77e2b9e4c869a1bf54ad0d20982235c5f5d2c5fb00eb31e1263727092cd1 not found: ID does not exist" Nov 24 17:39:43 crc kubenswrapper[4808]: I1124 17:39:43.203071 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-kn28w"] Nov 24 17:39:43 crc kubenswrapper[4808]: I1124 17:39:43.208552 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-kn28w"] Nov 24 17:39:44 crc kubenswrapper[4808]: I1124 17:39:44.356335 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="666531cc-9f5f-4f42-8c28-e3de1f361257" path="/var/lib/kubelet/pods/666531cc-9f5f-4f42-8c28-e3de1f361257/volumes" Nov 24 17:39:46 crc kubenswrapper[4808]: I1124 17:39:46.189482 4808 generic.go:334] "Generic (PLEG): container finished" podID="74d302c4-75f5-4cd4-a72e-4c76961320c7" containerID="3ccf5aaf604e770a7a7378c374474ee8e6772c72fdd1ac3dace809dd1d118d39" exitCode=0 Nov 24 17:39:46 crc kubenswrapper[4808]: I1124 17:39:46.189542 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wdtd" event={"ID":"74d302c4-75f5-4cd4-a72e-4c76961320c7","Type":"ContainerDied","Data":"3ccf5aaf604e770a7a7378c374474ee8e6772c72fdd1ac3dace809dd1d118d39"} Nov 24 17:39:48 crc kubenswrapper[4808]: I1124 17:39:48.038139 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-d6xmw" Nov 24 17:39:48 crc kubenswrapper[4808]: I1124 17:39:48.038573 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-d6xmw" Nov 24 17:39:48 crc kubenswrapper[4808]: I1124 17:39:48.069192 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-d6xmw" Nov 24 17:39:48 crc kubenswrapper[4808]: I1124 17:39:48.205256 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wdtd" event={"ID":"74d302c4-75f5-4cd4-a72e-4c76961320c7","Type":"ContainerStarted","Data":"5e260443eeb5647f65073059df9b5f2730aac3ba1ee57ae27f3e9d5e52d8ad5a"} Nov 24 17:39:48 crc kubenswrapper[4808]: I1124 17:39:48.229760 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9wdtd" podStartSLOduration=3.231921648 podStartE2EDuration="7.229737764s" podCreationTimestamp="2025-11-24 17:39:41 +0000 UTC" firstStartedPulling="2025-11-24 17:39:43.158593692 +0000 UTC m=+775.756261484" lastFinishedPulling="2025-11-24 17:39:47.156409808 +0000 UTC m=+779.754077600" observedRunningTime="2025-11-24 17:39:48.222748692 +0000 UTC m=+780.820416484" watchObservedRunningTime="2025-11-24 17:39:48.229737764 +0000 UTC m=+780.827405576" Nov 24 17:39:48 crc kubenswrapper[4808]: I1124 17:39:48.235494 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-d6xmw" Nov 24 17:39:49 crc kubenswrapper[4808]: I1124 17:39:49.226973 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-b46nx" Nov 24 17:39:50 crc kubenswrapper[4808]: I1124 17:39:50.949549 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6"] Nov 24 17:39:50 crc kubenswrapper[4808]: E1124 17:39:50.949887 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="666531cc-9f5f-4f42-8c28-e3de1f361257" containerName="registry-server" Nov 24 17:39:50 crc kubenswrapper[4808]: I1124 17:39:50.949904 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="666531cc-9f5f-4f42-8c28-e3de1f361257" containerName="registry-server" Nov 24 17:39:50 crc kubenswrapper[4808]: I1124 17:39:50.950074 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="666531cc-9f5f-4f42-8c28-e3de1f361257" containerName="registry-server" Nov 24 17:39:50 crc kubenswrapper[4808]: I1124 17:39:50.951131 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" Nov 24 17:39:50 crc kubenswrapper[4808]: I1124 17:39:50.953347 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-cfp59" Nov 24 17:39:50 crc kubenswrapper[4808]: I1124 17:39:50.965383 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6"] Nov 24 17:39:51 crc kubenswrapper[4808]: I1124 17:39:51.080374 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/eb16e3fe-9175-449c-af90-76b3d2475753-bundle\") pod \"f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6\" (UID: \"eb16e3fe-9175-449c-af90-76b3d2475753\") " pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" Nov 24 17:39:51 crc kubenswrapper[4808]: I1124 17:39:51.080745 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/eb16e3fe-9175-449c-af90-76b3d2475753-util\") pod \"f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6\" (UID: \"eb16e3fe-9175-449c-af90-76b3d2475753\") " pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" Nov 24 17:39:51 crc kubenswrapper[4808]: I1124 17:39:51.080804 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2jgc\" (UniqueName: \"kubernetes.io/projected/eb16e3fe-9175-449c-af90-76b3d2475753-kube-api-access-f2jgc\") pod \"f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6\" (UID: \"eb16e3fe-9175-449c-af90-76b3d2475753\") " pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" Nov 24 17:39:51 crc kubenswrapper[4808]: I1124 17:39:51.181799 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2jgc\" (UniqueName: \"kubernetes.io/projected/eb16e3fe-9175-449c-af90-76b3d2475753-kube-api-access-f2jgc\") pod \"f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6\" (UID: \"eb16e3fe-9175-449c-af90-76b3d2475753\") " pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" Nov 24 17:39:51 crc kubenswrapper[4808]: I1124 17:39:51.181923 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/eb16e3fe-9175-449c-af90-76b3d2475753-bundle\") pod \"f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6\" (UID: \"eb16e3fe-9175-449c-af90-76b3d2475753\") " pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" Nov 24 17:39:51 crc kubenswrapper[4808]: I1124 17:39:51.181970 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/eb16e3fe-9175-449c-af90-76b3d2475753-util\") pod \"f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6\" (UID: \"eb16e3fe-9175-449c-af90-76b3d2475753\") " pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" Nov 24 17:39:51 crc kubenswrapper[4808]: I1124 17:39:51.182517 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/eb16e3fe-9175-449c-af90-76b3d2475753-util\") pod \"f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6\" (UID: \"eb16e3fe-9175-449c-af90-76b3d2475753\") " pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" Nov 24 17:39:51 crc kubenswrapper[4808]: I1124 17:39:51.183110 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/eb16e3fe-9175-449c-af90-76b3d2475753-bundle\") pod \"f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6\" (UID: \"eb16e3fe-9175-449c-af90-76b3d2475753\") " pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" Nov 24 17:39:51 crc kubenswrapper[4808]: I1124 17:39:51.203729 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2jgc\" (UniqueName: \"kubernetes.io/projected/eb16e3fe-9175-449c-af90-76b3d2475753-kube-api-access-f2jgc\") pod \"f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6\" (UID: \"eb16e3fe-9175-449c-af90-76b3d2475753\") " pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" Nov 24 17:39:51 crc kubenswrapper[4808]: I1124 17:39:51.279560 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" Nov 24 17:39:51 crc kubenswrapper[4808]: I1124 17:39:51.735115 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6"] Nov 24 17:39:51 crc kubenswrapper[4808]: W1124 17:39:51.749971 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb16e3fe_9175_449c_af90_76b3d2475753.slice/crio-5895973d1b9a605712ea73e72cc9ae7b7378a1d76cd70cab88d9db271f71223a WatchSource:0}: Error finding container 5895973d1b9a605712ea73e72cc9ae7b7378a1d76cd70cab88d9db271f71223a: Status 404 returned error can't find the container with id 5895973d1b9a605712ea73e72cc9ae7b7378a1d76cd70cab88d9db271f71223a Nov 24 17:39:51 crc kubenswrapper[4808]: I1124 17:39:51.845791 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:51 crc kubenswrapper[4808]: I1124 17:39:51.845911 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:51 crc kubenswrapper[4808]: I1124 17:39:51.895980 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:52 crc kubenswrapper[4808]: I1124 17:39:52.233704 4808 generic.go:334] "Generic (PLEG): container finished" podID="eb16e3fe-9175-449c-af90-76b3d2475753" containerID="affdfce6348e5bffc7985892d50651a14be5b4758f109992f4746be9659251f8" exitCode=0 Nov 24 17:39:52 crc kubenswrapper[4808]: I1124 17:39:52.233851 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" event={"ID":"eb16e3fe-9175-449c-af90-76b3d2475753","Type":"ContainerDied","Data":"affdfce6348e5bffc7985892d50651a14be5b4758f109992f4746be9659251f8"} Nov 24 17:39:52 crc kubenswrapper[4808]: I1124 17:39:52.233905 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" event={"ID":"eb16e3fe-9175-449c-af90-76b3d2475753","Type":"ContainerStarted","Data":"5895973d1b9a605712ea73e72cc9ae7b7378a1d76cd70cab88d9db271f71223a"} Nov 24 17:39:52 crc kubenswrapper[4808]: I1124 17:39:52.280205 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:53 crc kubenswrapper[4808]: I1124 17:39:53.244554 4808 generic.go:334] "Generic (PLEG): container finished" podID="eb16e3fe-9175-449c-af90-76b3d2475753" containerID="c91a9bf6a3036572114481eda282c1e1a004fb6b1b8f2a86306757c4b2110468" exitCode=0 Nov 24 17:39:53 crc kubenswrapper[4808]: I1124 17:39:53.244683 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" event={"ID":"eb16e3fe-9175-449c-af90-76b3d2475753","Type":"ContainerDied","Data":"c91a9bf6a3036572114481eda282c1e1a004fb6b1b8f2a86306757c4b2110468"} Nov 24 17:39:53 crc kubenswrapper[4808]: I1124 17:39:53.300053 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9wdtd"] Nov 24 17:39:54 crc kubenswrapper[4808]: I1124 17:39:54.254082 4808 generic.go:334] "Generic (PLEG): container finished" podID="eb16e3fe-9175-449c-af90-76b3d2475753" containerID="3565a0cee7f2f36e16dff170c78079b8167e4412ac5249517d64f5cf041c1ab8" exitCode=0 Nov 24 17:39:54 crc kubenswrapper[4808]: I1124 17:39:54.254237 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" event={"ID":"eb16e3fe-9175-449c-af90-76b3d2475753","Type":"ContainerDied","Data":"3565a0cee7f2f36e16dff170c78079b8167e4412ac5249517d64f5cf041c1ab8"} Nov 24 17:39:54 crc kubenswrapper[4808]: I1124 17:39:54.297904 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7dfh8"] Nov 24 17:39:54 crc kubenswrapper[4808]: I1124 17:39:54.299602 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:39:54 crc kubenswrapper[4808]: I1124 17:39:54.308462 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7dfh8"] Nov 24 17:39:54 crc kubenswrapper[4808]: I1124 17:39:54.430538 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmjl8\" (UniqueName: \"kubernetes.io/projected/15731e66-4629-4f34-af6b-83ae11044e5f-kube-api-access-jmjl8\") pod \"certified-operators-7dfh8\" (UID: \"15731e66-4629-4f34-af6b-83ae11044e5f\") " pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:39:54 crc kubenswrapper[4808]: I1124 17:39:54.430648 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15731e66-4629-4f34-af6b-83ae11044e5f-utilities\") pod \"certified-operators-7dfh8\" (UID: \"15731e66-4629-4f34-af6b-83ae11044e5f\") " pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:39:54 crc kubenswrapper[4808]: I1124 17:39:54.431062 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15731e66-4629-4f34-af6b-83ae11044e5f-catalog-content\") pod \"certified-operators-7dfh8\" (UID: \"15731e66-4629-4f34-af6b-83ae11044e5f\") " pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:39:54 crc kubenswrapper[4808]: I1124 17:39:54.532483 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15731e66-4629-4f34-af6b-83ae11044e5f-catalog-content\") pod \"certified-operators-7dfh8\" (UID: \"15731e66-4629-4f34-af6b-83ae11044e5f\") " pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:39:54 crc kubenswrapper[4808]: I1124 17:39:54.532592 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmjl8\" (UniqueName: \"kubernetes.io/projected/15731e66-4629-4f34-af6b-83ae11044e5f-kube-api-access-jmjl8\") pod \"certified-operators-7dfh8\" (UID: \"15731e66-4629-4f34-af6b-83ae11044e5f\") " pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:39:54 crc kubenswrapper[4808]: I1124 17:39:54.532650 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15731e66-4629-4f34-af6b-83ae11044e5f-utilities\") pod \"certified-operators-7dfh8\" (UID: \"15731e66-4629-4f34-af6b-83ae11044e5f\") " pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:39:54 crc kubenswrapper[4808]: I1124 17:39:54.533186 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15731e66-4629-4f34-af6b-83ae11044e5f-catalog-content\") pod \"certified-operators-7dfh8\" (UID: \"15731e66-4629-4f34-af6b-83ae11044e5f\") " pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:39:54 crc kubenswrapper[4808]: I1124 17:39:54.533252 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15731e66-4629-4f34-af6b-83ae11044e5f-utilities\") pod \"certified-operators-7dfh8\" (UID: \"15731e66-4629-4f34-af6b-83ae11044e5f\") " pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:39:54 crc kubenswrapper[4808]: I1124 17:39:54.556579 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmjl8\" (UniqueName: \"kubernetes.io/projected/15731e66-4629-4f34-af6b-83ae11044e5f-kube-api-access-jmjl8\") pod \"certified-operators-7dfh8\" (UID: \"15731e66-4629-4f34-af6b-83ae11044e5f\") " pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:39:54 crc kubenswrapper[4808]: I1124 17:39:54.616672 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.088587 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7dfh8"] Nov 24 17:39:55 crc kubenswrapper[4808]: W1124 17:39:55.100188 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15731e66_4629_4f34_af6b_83ae11044e5f.slice/crio-9a0e93c96b132d08a489dd787ca48d54142d1c7e14362ea4dcad0971bccccc48 WatchSource:0}: Error finding container 9a0e93c96b132d08a489dd787ca48d54142d1c7e14362ea4dcad0971bccccc48: Status 404 returned error can't find the container with id 9a0e93c96b132d08a489dd787ca48d54142d1c7e14362ea4dcad0971bccccc48 Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.263136 4808 generic.go:334] "Generic (PLEG): container finished" podID="15731e66-4629-4f34-af6b-83ae11044e5f" containerID="b15adb496bf3b395b1bebf26f3acb10ba88ac529dcbc7c370ea8e99cb1d1b78e" exitCode=0 Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.263303 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dfh8" event={"ID":"15731e66-4629-4f34-af6b-83ae11044e5f","Type":"ContainerDied","Data":"b15adb496bf3b395b1bebf26f3acb10ba88ac529dcbc7c370ea8e99cb1d1b78e"} Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.263596 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dfh8" event={"ID":"15731e66-4629-4f34-af6b-83ae11044e5f","Type":"ContainerStarted","Data":"9a0e93c96b132d08a489dd787ca48d54142d1c7e14362ea4dcad0971bccccc48"} Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.264236 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9wdtd" podUID="74d302c4-75f5-4cd4-a72e-4c76961320c7" containerName="registry-server" containerID="cri-o://5e260443eeb5647f65073059df9b5f2730aac3ba1ee57ae27f3e9d5e52d8ad5a" gracePeriod=2 Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.516295 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.647307 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/eb16e3fe-9175-449c-af90-76b3d2475753-bundle\") pod \"eb16e3fe-9175-449c-af90-76b3d2475753\" (UID: \"eb16e3fe-9175-449c-af90-76b3d2475753\") " Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.647426 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2jgc\" (UniqueName: \"kubernetes.io/projected/eb16e3fe-9175-449c-af90-76b3d2475753-kube-api-access-f2jgc\") pod \"eb16e3fe-9175-449c-af90-76b3d2475753\" (UID: \"eb16e3fe-9175-449c-af90-76b3d2475753\") " Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.647532 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/eb16e3fe-9175-449c-af90-76b3d2475753-util\") pod \"eb16e3fe-9175-449c-af90-76b3d2475753\" (UID: \"eb16e3fe-9175-449c-af90-76b3d2475753\") " Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.648124 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb16e3fe-9175-449c-af90-76b3d2475753-bundle" (OuterVolumeSpecName: "bundle") pod "eb16e3fe-9175-449c-af90-76b3d2475753" (UID: "eb16e3fe-9175-449c-af90-76b3d2475753"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.654193 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb16e3fe-9175-449c-af90-76b3d2475753-kube-api-access-f2jgc" (OuterVolumeSpecName: "kube-api-access-f2jgc") pod "eb16e3fe-9175-449c-af90-76b3d2475753" (UID: "eb16e3fe-9175-449c-af90-76b3d2475753"). InnerVolumeSpecName "kube-api-access-f2jgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.663424 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb16e3fe-9175-449c-af90-76b3d2475753-util" (OuterVolumeSpecName: "util") pod "eb16e3fe-9175-449c-af90-76b3d2475753" (UID: "eb16e3fe-9175-449c-af90-76b3d2475753"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.689096 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.748979 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbb59\" (UniqueName: \"kubernetes.io/projected/74d302c4-75f5-4cd4-a72e-4c76961320c7-kube-api-access-sbb59\") pod \"74d302c4-75f5-4cd4-a72e-4c76961320c7\" (UID: \"74d302c4-75f5-4cd4-a72e-4c76961320c7\") " Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.749188 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74d302c4-75f5-4cd4-a72e-4c76961320c7-utilities\") pod \"74d302c4-75f5-4cd4-a72e-4c76961320c7\" (UID: \"74d302c4-75f5-4cd4-a72e-4c76961320c7\") " Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.749333 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74d302c4-75f5-4cd4-a72e-4c76961320c7-catalog-content\") pod \"74d302c4-75f5-4cd4-a72e-4c76961320c7\" (UID: \"74d302c4-75f5-4cd4-a72e-4c76961320c7\") " Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.749686 4808 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/eb16e3fe-9175-449c-af90-76b3d2475753-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.749707 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2jgc\" (UniqueName: \"kubernetes.io/projected/eb16e3fe-9175-449c-af90-76b3d2475753-kube-api-access-f2jgc\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.749720 4808 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/eb16e3fe-9175-449c-af90-76b3d2475753-util\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.750363 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74d302c4-75f5-4cd4-a72e-4c76961320c7-utilities" (OuterVolumeSpecName: "utilities") pod "74d302c4-75f5-4cd4-a72e-4c76961320c7" (UID: "74d302c4-75f5-4cd4-a72e-4c76961320c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.756226 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74d302c4-75f5-4cd4-a72e-4c76961320c7-kube-api-access-sbb59" (OuterVolumeSpecName: "kube-api-access-sbb59") pod "74d302c4-75f5-4cd4-a72e-4c76961320c7" (UID: "74d302c4-75f5-4cd4-a72e-4c76961320c7"). InnerVolumeSpecName "kube-api-access-sbb59". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.803350 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74d302c4-75f5-4cd4-a72e-4c76961320c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74d302c4-75f5-4cd4-a72e-4c76961320c7" (UID: "74d302c4-75f5-4cd4-a72e-4c76961320c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.851103 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbb59\" (UniqueName: \"kubernetes.io/projected/74d302c4-75f5-4cd4-a72e-4c76961320c7-kube-api-access-sbb59\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.851150 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74d302c4-75f5-4cd4-a72e-4c76961320c7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:55 crc kubenswrapper[4808]: I1124 17:39:55.851165 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74d302c4-75f5-4cd4-a72e-4c76961320c7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.273983 4808 generic.go:334] "Generic (PLEG): container finished" podID="15731e66-4629-4f34-af6b-83ae11044e5f" containerID="f83a611a055404d87ae624dfdd727b73558c2e6e48d318af10eade3576ad74df" exitCode=0 Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.274065 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dfh8" event={"ID":"15731e66-4629-4f34-af6b-83ae11044e5f","Type":"ContainerDied","Data":"f83a611a055404d87ae624dfdd727b73558c2e6e48d318af10eade3576ad74df"} Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.277890 4808 generic.go:334] "Generic (PLEG): container finished" podID="74d302c4-75f5-4cd4-a72e-4c76961320c7" containerID="5e260443eeb5647f65073059df9b5f2730aac3ba1ee57ae27f3e9d5e52d8ad5a" exitCode=0 Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.277981 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wdtd" event={"ID":"74d302c4-75f5-4cd4-a72e-4c76961320c7","Type":"ContainerDied","Data":"5e260443eeb5647f65073059df9b5f2730aac3ba1ee57ae27f3e9d5e52d8ad5a"} Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.278040 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wdtd" event={"ID":"74d302c4-75f5-4cd4-a72e-4c76961320c7","Type":"ContainerDied","Data":"8d44aef9ee6ac165b931cae7aa0bba5184c759b39789cd8bee9dc36cb3087c18"} Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.278065 4808 scope.go:117] "RemoveContainer" containerID="5e260443eeb5647f65073059df9b5f2730aac3ba1ee57ae27f3e9d5e52d8ad5a" Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.278382 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wdtd" Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.282045 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" event={"ID":"eb16e3fe-9175-449c-af90-76b3d2475753","Type":"ContainerDied","Data":"5895973d1b9a605712ea73e72cc9ae7b7378a1d76cd70cab88d9db271f71223a"} Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.282074 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5895973d1b9a605712ea73e72cc9ae7b7378a1d76cd70cab88d9db271f71223a" Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.282149 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6" Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.301448 4808 scope.go:117] "RemoveContainer" containerID="3ccf5aaf604e770a7a7378c374474ee8e6772c72fdd1ac3dace809dd1d118d39" Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.328169 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9wdtd"] Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.335107 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9wdtd"] Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.338374 4808 scope.go:117] "RemoveContainer" containerID="9dda9bf93dd2c6f27fe90c8907e6ac375539628e44e075d7de3f5f9913bdea00" Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.357394 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74d302c4-75f5-4cd4-a72e-4c76961320c7" path="/var/lib/kubelet/pods/74d302c4-75f5-4cd4-a72e-4c76961320c7/volumes" Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.357419 4808 scope.go:117] "RemoveContainer" containerID="5e260443eeb5647f65073059df9b5f2730aac3ba1ee57ae27f3e9d5e52d8ad5a" Nov 24 17:39:56 crc kubenswrapper[4808]: E1124 17:39:56.357901 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e260443eeb5647f65073059df9b5f2730aac3ba1ee57ae27f3e9d5e52d8ad5a\": container with ID starting with 5e260443eeb5647f65073059df9b5f2730aac3ba1ee57ae27f3e9d5e52d8ad5a not found: ID does not exist" containerID="5e260443eeb5647f65073059df9b5f2730aac3ba1ee57ae27f3e9d5e52d8ad5a" Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.357933 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e260443eeb5647f65073059df9b5f2730aac3ba1ee57ae27f3e9d5e52d8ad5a"} err="failed to get container status \"5e260443eeb5647f65073059df9b5f2730aac3ba1ee57ae27f3e9d5e52d8ad5a\": rpc error: code = NotFound desc = could not find container \"5e260443eeb5647f65073059df9b5f2730aac3ba1ee57ae27f3e9d5e52d8ad5a\": container with ID starting with 5e260443eeb5647f65073059df9b5f2730aac3ba1ee57ae27f3e9d5e52d8ad5a not found: ID does not exist" Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.358000 4808 scope.go:117] "RemoveContainer" containerID="3ccf5aaf604e770a7a7378c374474ee8e6772c72fdd1ac3dace809dd1d118d39" Nov 24 17:39:56 crc kubenswrapper[4808]: E1124 17:39:56.358330 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ccf5aaf604e770a7a7378c374474ee8e6772c72fdd1ac3dace809dd1d118d39\": container with ID starting with 3ccf5aaf604e770a7a7378c374474ee8e6772c72fdd1ac3dace809dd1d118d39 not found: ID does not exist" containerID="3ccf5aaf604e770a7a7378c374474ee8e6772c72fdd1ac3dace809dd1d118d39" Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.358368 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ccf5aaf604e770a7a7378c374474ee8e6772c72fdd1ac3dace809dd1d118d39"} err="failed to get container status \"3ccf5aaf604e770a7a7378c374474ee8e6772c72fdd1ac3dace809dd1d118d39\": rpc error: code = NotFound desc = could not find container \"3ccf5aaf604e770a7a7378c374474ee8e6772c72fdd1ac3dace809dd1d118d39\": container with ID starting with 3ccf5aaf604e770a7a7378c374474ee8e6772c72fdd1ac3dace809dd1d118d39 not found: ID does not exist" Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.358395 4808 scope.go:117] "RemoveContainer" containerID="9dda9bf93dd2c6f27fe90c8907e6ac375539628e44e075d7de3f5f9913bdea00" Nov 24 17:39:56 crc kubenswrapper[4808]: E1124 17:39:56.358726 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dda9bf93dd2c6f27fe90c8907e6ac375539628e44e075d7de3f5f9913bdea00\": container with ID starting with 9dda9bf93dd2c6f27fe90c8907e6ac375539628e44e075d7de3f5f9913bdea00 not found: ID does not exist" containerID="9dda9bf93dd2c6f27fe90c8907e6ac375539628e44e075d7de3f5f9913bdea00" Nov 24 17:39:56 crc kubenswrapper[4808]: I1124 17:39:56.358752 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dda9bf93dd2c6f27fe90c8907e6ac375539628e44e075d7de3f5f9913bdea00"} err="failed to get container status \"9dda9bf93dd2c6f27fe90c8907e6ac375539628e44e075d7de3f5f9913bdea00\": rpc error: code = NotFound desc = could not find container \"9dda9bf93dd2c6f27fe90c8907e6ac375539628e44e075d7de3f5f9913bdea00\": container with ID starting with 9dda9bf93dd2c6f27fe90c8907e6ac375539628e44e075d7de3f5f9913bdea00 not found: ID does not exist" Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.292157 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dfh8" event={"ID":"15731e66-4629-4f34-af6b-83ae11044e5f","Type":"ContainerStarted","Data":"3f0bfb9401fa772545bbfd13fc2416f5321ce1e00a29c01336a605b897858ce8"} Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.311711 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7dfh8" podStartSLOduration=1.8959035480000002 podStartE2EDuration="3.311692598s" podCreationTimestamp="2025-11-24 17:39:54 +0000 UTC" firstStartedPulling="2025-11-24 17:39:55.264860789 +0000 UTC m=+787.862528591" lastFinishedPulling="2025-11-24 17:39:56.680649839 +0000 UTC m=+789.278317641" observedRunningTime="2025-11-24 17:39:57.308715676 +0000 UTC m=+789.906383478" watchObservedRunningTime="2025-11-24 17:39:57.311692598 +0000 UTC m=+789.909360420" Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.898061 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-25g4f"] Nov 24 17:39:57 crc kubenswrapper[4808]: E1124 17:39:57.898360 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74d302c4-75f5-4cd4-a72e-4c76961320c7" containerName="registry-server" Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.898374 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="74d302c4-75f5-4cd4-a72e-4c76961320c7" containerName="registry-server" Nov 24 17:39:57 crc kubenswrapper[4808]: E1124 17:39:57.898386 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74d302c4-75f5-4cd4-a72e-4c76961320c7" containerName="extract-utilities" Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.898396 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="74d302c4-75f5-4cd4-a72e-4c76961320c7" containerName="extract-utilities" Nov 24 17:39:57 crc kubenswrapper[4808]: E1124 17:39:57.898410 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb16e3fe-9175-449c-af90-76b3d2475753" containerName="util" Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.898419 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb16e3fe-9175-449c-af90-76b3d2475753" containerName="util" Nov 24 17:39:57 crc kubenswrapper[4808]: E1124 17:39:57.898429 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb16e3fe-9175-449c-af90-76b3d2475753" containerName="pull" Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.898436 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb16e3fe-9175-449c-af90-76b3d2475753" containerName="pull" Nov 24 17:39:57 crc kubenswrapper[4808]: E1124 17:39:57.898455 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74d302c4-75f5-4cd4-a72e-4c76961320c7" containerName="extract-content" Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.898461 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="74d302c4-75f5-4cd4-a72e-4c76961320c7" containerName="extract-content" Nov 24 17:39:57 crc kubenswrapper[4808]: E1124 17:39:57.898470 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb16e3fe-9175-449c-af90-76b3d2475753" containerName="extract" Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.898477 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb16e3fe-9175-449c-af90-76b3d2475753" containerName="extract" Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.898600 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb16e3fe-9175-449c-af90-76b3d2475753" containerName="extract" Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.898614 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="74d302c4-75f5-4cd4-a72e-4c76961320c7" containerName="registry-server" Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.899671 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.914336 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-25g4f"] Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.980917 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mmsx\" (UniqueName: \"kubernetes.io/projected/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-kube-api-access-8mmsx\") pod \"redhat-marketplace-25g4f\" (UID: \"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a\") " pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.981006 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-utilities\") pod \"redhat-marketplace-25g4f\" (UID: \"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a\") " pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:39:57 crc kubenswrapper[4808]: I1124 17:39:57.981062 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-catalog-content\") pod \"redhat-marketplace-25g4f\" (UID: \"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a\") " pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:39:58 crc kubenswrapper[4808]: I1124 17:39:58.082757 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mmsx\" (UniqueName: \"kubernetes.io/projected/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-kube-api-access-8mmsx\") pod \"redhat-marketplace-25g4f\" (UID: \"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a\") " pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:39:58 crc kubenswrapper[4808]: I1124 17:39:58.082861 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-utilities\") pod \"redhat-marketplace-25g4f\" (UID: \"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a\") " pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:39:58 crc kubenswrapper[4808]: I1124 17:39:58.082904 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-catalog-content\") pod \"redhat-marketplace-25g4f\" (UID: \"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a\") " pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:39:58 crc kubenswrapper[4808]: I1124 17:39:58.083894 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-catalog-content\") pod \"redhat-marketplace-25g4f\" (UID: \"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a\") " pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:39:58 crc kubenswrapper[4808]: I1124 17:39:58.084545 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-utilities\") pod \"redhat-marketplace-25g4f\" (UID: \"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a\") " pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:39:58 crc kubenswrapper[4808]: I1124 17:39:58.117550 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mmsx\" (UniqueName: \"kubernetes.io/projected/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-kube-api-access-8mmsx\") pod \"redhat-marketplace-25g4f\" (UID: \"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a\") " pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:39:58 crc kubenswrapper[4808]: I1124 17:39:58.216192 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:39:58 crc kubenswrapper[4808]: I1124 17:39:58.660959 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-25g4f"] Nov 24 17:39:59 crc kubenswrapper[4808]: I1124 17:39:59.307937 4808 generic.go:334] "Generic (PLEG): container finished" podID="7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a" containerID="e05653b968a2ce578b889af8539af7671e38988e181f4ccaaad521a67ca188e9" exitCode=0 Nov 24 17:39:59 crc kubenswrapper[4808]: I1124 17:39:59.308006 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25g4f" event={"ID":"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a","Type":"ContainerDied","Data":"e05653b968a2ce578b889af8539af7671e38988e181f4ccaaad521a67ca188e9"} Nov 24 17:39:59 crc kubenswrapper[4808]: I1124 17:39:59.308066 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25g4f" event={"ID":"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a","Type":"ContainerStarted","Data":"19992ea41273cb08aa9376835ef34ec688a2b99f0083204a0b072c1354ac7179"} Nov 24 17:40:01 crc kubenswrapper[4808]: I1124 17:40:01.390968 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-798b7b4bc9-dhdfm"] Nov 24 17:40:01 crc kubenswrapper[4808]: I1124 17:40:01.392337 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-798b7b4bc9-dhdfm" Nov 24 17:40:01 crc kubenswrapper[4808]: I1124 17:40:01.394381 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-4d9dk" Nov 24 17:40:01 crc kubenswrapper[4808]: I1124 17:40:01.425095 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xprpv\" (UniqueName: \"kubernetes.io/projected/8d18f465-70ea-4d1f-8531-dcfee37c960d-kube-api-access-xprpv\") pod \"openstack-operator-controller-operator-798b7b4bc9-dhdfm\" (UID: \"8d18f465-70ea-4d1f-8531-dcfee37c960d\") " pod="openstack-operators/openstack-operator-controller-operator-798b7b4bc9-dhdfm" Nov 24 17:40:01 crc kubenswrapper[4808]: I1124 17:40:01.435739 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-798b7b4bc9-dhdfm"] Nov 24 17:40:01 crc kubenswrapper[4808]: I1124 17:40:01.526732 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xprpv\" (UniqueName: \"kubernetes.io/projected/8d18f465-70ea-4d1f-8531-dcfee37c960d-kube-api-access-xprpv\") pod \"openstack-operator-controller-operator-798b7b4bc9-dhdfm\" (UID: \"8d18f465-70ea-4d1f-8531-dcfee37c960d\") " pod="openstack-operators/openstack-operator-controller-operator-798b7b4bc9-dhdfm" Nov 24 17:40:01 crc kubenswrapper[4808]: I1124 17:40:01.551963 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xprpv\" (UniqueName: \"kubernetes.io/projected/8d18f465-70ea-4d1f-8531-dcfee37c960d-kube-api-access-xprpv\") pod \"openstack-operator-controller-operator-798b7b4bc9-dhdfm\" (UID: \"8d18f465-70ea-4d1f-8531-dcfee37c960d\") " pod="openstack-operators/openstack-operator-controller-operator-798b7b4bc9-dhdfm" Nov 24 17:40:01 crc kubenswrapper[4808]: I1124 17:40:01.713839 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-798b7b4bc9-dhdfm" Nov 24 17:40:02 crc kubenswrapper[4808]: I1124 17:40:02.362526 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25g4f" event={"ID":"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a","Type":"ContainerStarted","Data":"ff6594c8198122f8fd62db251337913e1e1ac71734b9e82db9537c2415f6b5b3"} Nov 24 17:40:02 crc kubenswrapper[4808]: I1124 17:40:02.420477 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-798b7b4bc9-dhdfm"] Nov 24 17:40:02 crc kubenswrapper[4808]: W1124 17:40:02.421098 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d18f465_70ea_4d1f_8531_dcfee37c960d.slice/crio-589f1b5ca2aee148526abe32d212226ed249fbf735c7608e848093bff5ca514b WatchSource:0}: Error finding container 589f1b5ca2aee148526abe32d212226ed249fbf735c7608e848093bff5ca514b: Status 404 returned error can't find the container with id 589f1b5ca2aee148526abe32d212226ed249fbf735c7608e848093bff5ca514b Nov 24 17:40:02 crc kubenswrapper[4808]: I1124 17:40:02.701316 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gc6db"] Nov 24 17:40:02 crc kubenswrapper[4808]: I1124 17:40:02.704205 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:02 crc kubenswrapper[4808]: I1124 17:40:02.714578 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gc6db"] Nov 24 17:40:02 crc kubenswrapper[4808]: I1124 17:40:02.751290 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrkcp\" (UniqueName: \"kubernetes.io/projected/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-kube-api-access-wrkcp\") pod \"redhat-operators-gc6db\" (UID: \"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2\") " pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:02 crc kubenswrapper[4808]: I1124 17:40:02.751364 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-catalog-content\") pod \"redhat-operators-gc6db\" (UID: \"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2\") " pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:02 crc kubenswrapper[4808]: I1124 17:40:02.751411 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-utilities\") pod \"redhat-operators-gc6db\" (UID: \"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2\") " pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:02 crc kubenswrapper[4808]: I1124 17:40:02.853431 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-catalog-content\") pod \"redhat-operators-gc6db\" (UID: \"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2\") " pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:02 crc kubenswrapper[4808]: I1124 17:40:02.853535 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-utilities\") pod \"redhat-operators-gc6db\" (UID: \"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2\") " pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:02 crc kubenswrapper[4808]: I1124 17:40:02.853625 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrkcp\" (UniqueName: \"kubernetes.io/projected/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-kube-api-access-wrkcp\") pod \"redhat-operators-gc6db\" (UID: \"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2\") " pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:02 crc kubenswrapper[4808]: I1124 17:40:02.854716 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-catalog-content\") pod \"redhat-operators-gc6db\" (UID: \"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2\") " pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:02 crc kubenswrapper[4808]: I1124 17:40:02.855037 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-utilities\") pod \"redhat-operators-gc6db\" (UID: \"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2\") " pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:02 crc kubenswrapper[4808]: I1124 17:40:02.885059 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrkcp\" (UniqueName: \"kubernetes.io/projected/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-kube-api-access-wrkcp\") pod \"redhat-operators-gc6db\" (UID: \"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2\") " pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:03 crc kubenswrapper[4808]: I1124 17:40:03.021757 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:03 crc kubenswrapper[4808]: I1124 17:40:03.323385 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gc6db"] Nov 24 17:40:03 crc kubenswrapper[4808]: W1124 17:40:03.335325 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e8a57e0_ef3f_4f8f_ae13_70b05ca9f9c2.slice/crio-d107e89c1902d2a85fe1046990fb93153a61dada55211ec47ced8777b42826ad WatchSource:0}: Error finding container d107e89c1902d2a85fe1046990fb93153a61dada55211ec47ced8777b42826ad: Status 404 returned error can't find the container with id d107e89c1902d2a85fe1046990fb93153a61dada55211ec47ced8777b42826ad Nov 24 17:40:03 crc kubenswrapper[4808]: I1124 17:40:03.361988 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gc6db" event={"ID":"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2","Type":"ContainerStarted","Data":"d107e89c1902d2a85fe1046990fb93153a61dada55211ec47ced8777b42826ad"} Nov 24 17:40:03 crc kubenswrapper[4808]: I1124 17:40:03.365239 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-798b7b4bc9-dhdfm" event={"ID":"8d18f465-70ea-4d1f-8531-dcfee37c960d","Type":"ContainerStarted","Data":"589f1b5ca2aee148526abe32d212226ed249fbf735c7608e848093bff5ca514b"} Nov 24 17:40:03 crc kubenswrapper[4808]: I1124 17:40:03.368057 4808 generic.go:334] "Generic (PLEG): container finished" podID="7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a" containerID="ff6594c8198122f8fd62db251337913e1e1ac71734b9e82db9537c2415f6b5b3" exitCode=0 Nov 24 17:40:03 crc kubenswrapper[4808]: I1124 17:40:03.368096 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25g4f" event={"ID":"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a","Type":"ContainerDied","Data":"ff6594c8198122f8fd62db251337913e1e1ac71734b9e82db9537c2415f6b5b3"} Nov 24 17:40:04 crc kubenswrapper[4808]: I1124 17:40:04.380372 4808 generic.go:334] "Generic (PLEG): container finished" podID="8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2" containerID="20594e11b9fae81a8a39584f84ec9aa4b83bb6930e037c580a93728b43f8e251" exitCode=0 Nov 24 17:40:04 crc kubenswrapper[4808]: I1124 17:40:04.380597 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gc6db" event={"ID":"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2","Type":"ContainerDied","Data":"20594e11b9fae81a8a39584f84ec9aa4b83bb6930e037c580a93728b43f8e251"} Nov 24 17:40:04 crc kubenswrapper[4808]: I1124 17:40:04.617665 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:40:04 crc kubenswrapper[4808]: I1124 17:40:04.618058 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:40:04 crc kubenswrapper[4808]: I1124 17:40:04.666795 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:40:05 crc kubenswrapper[4808]: I1124 17:40:05.444818 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:40:06 crc kubenswrapper[4808]: I1124 17:40:06.522559 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:40:06 crc kubenswrapper[4808]: I1124 17:40:06.522620 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:40:09 crc kubenswrapper[4808]: I1124 17:40:09.689471 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7dfh8"] Nov 24 17:40:09 crc kubenswrapper[4808]: I1124 17:40:09.690035 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7dfh8" podUID="15731e66-4629-4f34-af6b-83ae11044e5f" containerName="registry-server" containerID="cri-o://3f0bfb9401fa772545bbfd13fc2416f5321ce1e00a29c01336a605b897858ce8" gracePeriod=2 Nov 24 17:40:11 crc kubenswrapper[4808]: I1124 17:40:11.436227 4808 generic.go:334] "Generic (PLEG): container finished" podID="15731e66-4629-4f34-af6b-83ae11044e5f" containerID="3f0bfb9401fa772545bbfd13fc2416f5321ce1e00a29c01336a605b897858ce8" exitCode=0 Nov 24 17:40:11 crc kubenswrapper[4808]: I1124 17:40:11.436593 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dfh8" event={"ID":"15731e66-4629-4f34-af6b-83ae11044e5f","Type":"ContainerDied","Data":"3f0bfb9401fa772545bbfd13fc2416f5321ce1e00a29c01336a605b897858ce8"} Nov 24 17:40:11 crc kubenswrapper[4808]: I1124 17:40:11.995065 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.108738 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmjl8\" (UniqueName: \"kubernetes.io/projected/15731e66-4629-4f34-af6b-83ae11044e5f-kube-api-access-jmjl8\") pod \"15731e66-4629-4f34-af6b-83ae11044e5f\" (UID: \"15731e66-4629-4f34-af6b-83ae11044e5f\") " Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.109083 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15731e66-4629-4f34-af6b-83ae11044e5f-catalog-content\") pod \"15731e66-4629-4f34-af6b-83ae11044e5f\" (UID: \"15731e66-4629-4f34-af6b-83ae11044e5f\") " Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.109133 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15731e66-4629-4f34-af6b-83ae11044e5f-utilities\") pod \"15731e66-4629-4f34-af6b-83ae11044e5f\" (UID: \"15731e66-4629-4f34-af6b-83ae11044e5f\") " Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.110262 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15731e66-4629-4f34-af6b-83ae11044e5f-utilities" (OuterVolumeSpecName: "utilities") pod "15731e66-4629-4f34-af6b-83ae11044e5f" (UID: "15731e66-4629-4f34-af6b-83ae11044e5f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.116770 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15731e66-4629-4f34-af6b-83ae11044e5f-kube-api-access-jmjl8" (OuterVolumeSpecName: "kube-api-access-jmjl8") pod "15731e66-4629-4f34-af6b-83ae11044e5f" (UID: "15731e66-4629-4f34-af6b-83ae11044e5f"). InnerVolumeSpecName "kube-api-access-jmjl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.161953 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15731e66-4629-4f34-af6b-83ae11044e5f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15731e66-4629-4f34-af6b-83ae11044e5f" (UID: "15731e66-4629-4f34-af6b-83ae11044e5f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.210674 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmjl8\" (UniqueName: \"kubernetes.io/projected/15731e66-4629-4f34-af6b-83ae11044e5f-kube-api-access-jmjl8\") on node \"crc\" DevicePath \"\"" Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.210715 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15731e66-4629-4f34-af6b-83ae11044e5f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.210725 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15731e66-4629-4f34-af6b-83ae11044e5f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.445616 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-798b7b4bc9-dhdfm" event={"ID":"8d18f465-70ea-4d1f-8531-dcfee37c960d","Type":"ContainerStarted","Data":"3292385b711402e7469b51dc0dc6edc81f18c705d26504a320095cb55af20c68"} Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.446867 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-798b7b4bc9-dhdfm" Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.449240 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25g4f" event={"ID":"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a","Type":"ContainerStarted","Data":"0526e49b6bab870ac55976b8c4d540f155a64959b9430c48ca1bb98c744d49bd"} Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.456147 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dfh8" event={"ID":"15731e66-4629-4f34-af6b-83ae11044e5f","Type":"ContainerDied","Data":"9a0e93c96b132d08a489dd787ca48d54142d1c7e14362ea4dcad0971bccccc48"} Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.456222 4808 scope.go:117] "RemoveContainer" containerID="3f0bfb9401fa772545bbfd13fc2416f5321ce1e00a29c01336a605b897858ce8" Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.456416 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dfh8" Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.462298 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gc6db" event={"ID":"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2","Type":"ContainerStarted","Data":"f28a8e69ef57b9b7c639e4c1eaeaa76b528e468b65280cc01ce4c2dfd93d2b81"} Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.485871 4808 scope.go:117] "RemoveContainer" containerID="f83a611a055404d87ae624dfdd727b73558c2e6e48d318af10eade3576ad74df" Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.490125 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-798b7b4bc9-dhdfm" podStartSLOduration=1.9165203659999999 podStartE2EDuration="11.490105521s" podCreationTimestamp="2025-11-24 17:40:01 +0000 UTC" firstStartedPulling="2025-11-24 17:40:02.423585069 +0000 UTC m=+795.021252871" lastFinishedPulling="2025-11-24 17:40:11.997170224 +0000 UTC m=+804.594838026" observedRunningTime="2025-11-24 17:40:12.486258676 +0000 UTC m=+805.083926478" watchObservedRunningTime="2025-11-24 17:40:12.490105521 +0000 UTC m=+805.087773323" Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.511389 4808 scope.go:117] "RemoveContainer" containerID="b15adb496bf3b395b1bebf26f3acb10ba88ac529dcbc7c370ea8e99cb1d1b78e" Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.531228 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7dfh8"] Nov 24 17:40:12 crc kubenswrapper[4808]: I1124 17:40:12.535867 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7dfh8"] Nov 24 17:40:13 crc kubenswrapper[4808]: I1124 17:40:13.473802 4808 generic.go:334] "Generic (PLEG): container finished" podID="8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2" containerID="f28a8e69ef57b9b7c639e4c1eaeaa76b528e468b65280cc01ce4c2dfd93d2b81" exitCode=0 Nov 24 17:40:13 crc kubenswrapper[4808]: I1124 17:40:13.474288 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gc6db" event={"ID":"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2","Type":"ContainerDied","Data":"f28a8e69ef57b9b7c639e4c1eaeaa76b528e468b65280cc01ce4c2dfd93d2b81"} Nov 24 17:40:13 crc kubenswrapper[4808]: I1124 17:40:13.500223 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-25g4f" podStartSLOduration=4.48669266 podStartE2EDuration="16.50020427s" podCreationTimestamp="2025-11-24 17:39:57 +0000 UTC" firstStartedPulling="2025-11-24 17:39:59.310881829 +0000 UTC m=+791.908549631" lastFinishedPulling="2025-11-24 17:40:11.324393439 +0000 UTC m=+803.922061241" observedRunningTime="2025-11-24 17:40:12.558807958 +0000 UTC m=+805.156475760" watchObservedRunningTime="2025-11-24 17:40:13.50020427 +0000 UTC m=+806.097872092" Nov 24 17:40:14 crc kubenswrapper[4808]: I1124 17:40:14.356077 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15731e66-4629-4f34-af6b-83ae11044e5f" path="/var/lib/kubelet/pods/15731e66-4629-4f34-af6b-83ae11044e5f/volumes" Nov 24 17:40:14 crc kubenswrapper[4808]: I1124 17:40:14.485419 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gc6db" event={"ID":"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2","Type":"ContainerStarted","Data":"494679ee6a7db6c58f5489f38d54f969174bdf98f435414c1759971afdb117d6"} Nov 24 17:40:14 crc kubenswrapper[4808]: I1124 17:40:14.505066 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gc6db" podStartSLOduration=2.929431273 podStartE2EDuration="12.505044845s" podCreationTimestamp="2025-11-24 17:40:02 +0000 UTC" firstStartedPulling="2025-11-24 17:40:04.381916457 +0000 UTC m=+796.979584259" lastFinishedPulling="2025-11-24 17:40:13.957530029 +0000 UTC m=+806.555197831" observedRunningTime="2025-11-24 17:40:14.501771635 +0000 UTC m=+807.099439437" watchObservedRunningTime="2025-11-24 17:40:14.505044845 +0000 UTC m=+807.102712647" Nov 24 17:40:18 crc kubenswrapper[4808]: I1124 17:40:18.216720 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:40:18 crc kubenswrapper[4808]: I1124 17:40:18.217113 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:40:18 crc kubenswrapper[4808]: I1124 17:40:18.261638 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:40:18 crc kubenswrapper[4808]: I1124 17:40:18.554201 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:40:20 crc kubenswrapper[4808]: I1124 17:40:20.492091 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-25g4f"] Nov 24 17:40:20 crc kubenswrapper[4808]: I1124 17:40:20.527965 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-25g4f" podUID="7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a" containerName="registry-server" containerID="cri-o://0526e49b6bab870ac55976b8c4d540f155a64959b9430c48ca1bb98c744d49bd" gracePeriod=2 Nov 24 17:40:20 crc kubenswrapper[4808]: I1124 17:40:20.916739 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.037844 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mmsx\" (UniqueName: \"kubernetes.io/projected/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-kube-api-access-8mmsx\") pod \"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a\" (UID: \"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a\") " Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.038009 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-utilities\") pod \"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a\" (UID: \"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a\") " Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.038132 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-catalog-content\") pod \"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a\" (UID: \"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a\") " Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.039191 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-utilities" (OuterVolumeSpecName: "utilities") pod "7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a" (UID: "7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.044530 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-kube-api-access-8mmsx" (OuterVolumeSpecName: "kube-api-access-8mmsx") pod "7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a" (UID: "7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a"). InnerVolumeSpecName "kube-api-access-8mmsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.056486 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a" (UID: "7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.139831 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.139869 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.139885 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mmsx\" (UniqueName: \"kubernetes.io/projected/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a-kube-api-access-8mmsx\") on node \"crc\" DevicePath \"\"" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.537929 4808 generic.go:334] "Generic (PLEG): container finished" podID="7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a" containerID="0526e49b6bab870ac55976b8c4d540f155a64959b9430c48ca1bb98c744d49bd" exitCode=0 Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.537992 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25g4f" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.537992 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25g4f" event={"ID":"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a","Type":"ContainerDied","Data":"0526e49b6bab870ac55976b8c4d540f155a64959b9430c48ca1bb98c744d49bd"} Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.538491 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25g4f" event={"ID":"7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a","Type":"ContainerDied","Data":"19992ea41273cb08aa9376835ef34ec688a2b99f0083204a0b072c1354ac7179"} Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.538530 4808 scope.go:117] "RemoveContainer" containerID="0526e49b6bab870ac55976b8c4d540f155a64959b9430c48ca1bb98c744d49bd" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.566187 4808 scope.go:117] "RemoveContainer" containerID="ff6594c8198122f8fd62db251337913e1e1ac71734b9e82db9537c2415f6b5b3" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.577201 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-25g4f"] Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.580686 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-25g4f"] Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.591723 4808 scope.go:117] "RemoveContainer" containerID="e05653b968a2ce578b889af8539af7671e38988e181f4ccaaad521a67ca188e9" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.613748 4808 scope.go:117] "RemoveContainer" containerID="0526e49b6bab870ac55976b8c4d540f155a64959b9430c48ca1bb98c744d49bd" Nov 24 17:40:21 crc kubenswrapper[4808]: E1124 17:40:21.614365 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0526e49b6bab870ac55976b8c4d540f155a64959b9430c48ca1bb98c744d49bd\": container with ID starting with 0526e49b6bab870ac55976b8c4d540f155a64959b9430c48ca1bb98c744d49bd not found: ID does not exist" containerID="0526e49b6bab870ac55976b8c4d540f155a64959b9430c48ca1bb98c744d49bd" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.614416 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0526e49b6bab870ac55976b8c4d540f155a64959b9430c48ca1bb98c744d49bd"} err="failed to get container status \"0526e49b6bab870ac55976b8c4d540f155a64959b9430c48ca1bb98c744d49bd\": rpc error: code = NotFound desc = could not find container \"0526e49b6bab870ac55976b8c4d540f155a64959b9430c48ca1bb98c744d49bd\": container with ID starting with 0526e49b6bab870ac55976b8c4d540f155a64959b9430c48ca1bb98c744d49bd not found: ID does not exist" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.614447 4808 scope.go:117] "RemoveContainer" containerID="ff6594c8198122f8fd62db251337913e1e1ac71734b9e82db9537c2415f6b5b3" Nov 24 17:40:21 crc kubenswrapper[4808]: E1124 17:40:21.614864 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff6594c8198122f8fd62db251337913e1e1ac71734b9e82db9537c2415f6b5b3\": container with ID starting with ff6594c8198122f8fd62db251337913e1e1ac71734b9e82db9537c2415f6b5b3 not found: ID does not exist" containerID="ff6594c8198122f8fd62db251337913e1e1ac71734b9e82db9537c2415f6b5b3" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.614974 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff6594c8198122f8fd62db251337913e1e1ac71734b9e82db9537c2415f6b5b3"} err="failed to get container status \"ff6594c8198122f8fd62db251337913e1e1ac71734b9e82db9537c2415f6b5b3\": rpc error: code = NotFound desc = could not find container \"ff6594c8198122f8fd62db251337913e1e1ac71734b9e82db9537c2415f6b5b3\": container with ID starting with ff6594c8198122f8fd62db251337913e1e1ac71734b9e82db9537c2415f6b5b3 not found: ID does not exist" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.615004 4808 scope.go:117] "RemoveContainer" containerID="e05653b968a2ce578b889af8539af7671e38988e181f4ccaaad521a67ca188e9" Nov 24 17:40:21 crc kubenswrapper[4808]: E1124 17:40:21.615460 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e05653b968a2ce578b889af8539af7671e38988e181f4ccaaad521a67ca188e9\": container with ID starting with e05653b968a2ce578b889af8539af7671e38988e181f4ccaaad521a67ca188e9 not found: ID does not exist" containerID="e05653b968a2ce578b889af8539af7671e38988e181f4ccaaad521a67ca188e9" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.615486 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e05653b968a2ce578b889af8539af7671e38988e181f4ccaaad521a67ca188e9"} err="failed to get container status \"e05653b968a2ce578b889af8539af7671e38988e181f4ccaaad521a67ca188e9\": rpc error: code = NotFound desc = could not find container \"e05653b968a2ce578b889af8539af7671e38988e181f4ccaaad521a67ca188e9\": container with ID starting with e05653b968a2ce578b889af8539af7671e38988e181f4ccaaad521a67ca188e9 not found: ID does not exist" Nov 24 17:40:21 crc kubenswrapper[4808]: I1124 17:40:21.717225 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-798b7b4bc9-dhdfm" Nov 24 17:40:22 crc kubenswrapper[4808]: I1124 17:40:22.359980 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a" path="/var/lib/kubelet/pods/7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a/volumes" Nov 24 17:40:23 crc kubenswrapper[4808]: I1124 17:40:23.023031 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:23 crc kubenswrapper[4808]: I1124 17:40:23.023078 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:23 crc kubenswrapper[4808]: I1124 17:40:23.068824 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:23 crc kubenswrapper[4808]: I1124 17:40:23.595995 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:24 crc kubenswrapper[4808]: I1124 17:40:24.894393 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gc6db"] Nov 24 17:40:25 crc kubenswrapper[4808]: I1124 17:40:25.564803 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gc6db" podUID="8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2" containerName="registry-server" containerID="cri-o://494679ee6a7db6c58f5489f38d54f969174bdf98f435414c1759971afdb117d6" gracePeriod=2 Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.048469 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.110386 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-utilities\") pod \"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2\" (UID: \"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2\") " Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.110456 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrkcp\" (UniqueName: \"kubernetes.io/projected/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-kube-api-access-wrkcp\") pod \"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2\" (UID: \"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2\") " Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.110504 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-catalog-content\") pod \"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2\" (UID: \"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2\") " Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.124407 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-utilities" (OuterVolumeSpecName: "utilities") pod "8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2" (UID: "8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.132111 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-kube-api-access-wrkcp" (OuterVolumeSpecName: "kube-api-access-wrkcp") pod "8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2" (UID: "8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2"). InnerVolumeSpecName "kube-api-access-wrkcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.211912 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.211948 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrkcp\" (UniqueName: \"kubernetes.io/projected/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-kube-api-access-wrkcp\") on node \"crc\" DevicePath \"\"" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.247920 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2" (UID: "8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.314408 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.573452 4808 generic.go:334] "Generic (PLEG): container finished" podID="8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2" containerID="494679ee6a7db6c58f5489f38d54f969174bdf98f435414c1759971afdb117d6" exitCode=0 Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.573501 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gc6db" event={"ID":"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2","Type":"ContainerDied","Data":"494679ee6a7db6c58f5489f38d54f969174bdf98f435414c1759971afdb117d6"} Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.573541 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gc6db" event={"ID":"8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2","Type":"ContainerDied","Data":"d107e89c1902d2a85fe1046990fb93153a61dada55211ec47ced8777b42826ad"} Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.573544 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gc6db" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.573573 4808 scope.go:117] "RemoveContainer" containerID="494679ee6a7db6c58f5489f38d54f969174bdf98f435414c1759971afdb117d6" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.592831 4808 scope.go:117] "RemoveContainer" containerID="f28a8e69ef57b9b7c639e4c1eaeaa76b528e468b65280cc01ce4c2dfd93d2b81" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.602563 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gc6db"] Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.609745 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gc6db"] Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.611244 4808 scope.go:117] "RemoveContainer" containerID="20594e11b9fae81a8a39584f84ec9aa4b83bb6930e037c580a93728b43f8e251" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.646444 4808 scope.go:117] "RemoveContainer" containerID="494679ee6a7db6c58f5489f38d54f969174bdf98f435414c1759971afdb117d6" Nov 24 17:40:26 crc kubenswrapper[4808]: E1124 17:40:26.647484 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"494679ee6a7db6c58f5489f38d54f969174bdf98f435414c1759971afdb117d6\": container with ID starting with 494679ee6a7db6c58f5489f38d54f969174bdf98f435414c1759971afdb117d6 not found: ID does not exist" containerID="494679ee6a7db6c58f5489f38d54f969174bdf98f435414c1759971afdb117d6" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.647529 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"494679ee6a7db6c58f5489f38d54f969174bdf98f435414c1759971afdb117d6"} err="failed to get container status \"494679ee6a7db6c58f5489f38d54f969174bdf98f435414c1759971afdb117d6\": rpc error: code = NotFound desc = could not find container \"494679ee6a7db6c58f5489f38d54f969174bdf98f435414c1759971afdb117d6\": container with ID starting with 494679ee6a7db6c58f5489f38d54f969174bdf98f435414c1759971afdb117d6 not found: ID does not exist" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.647560 4808 scope.go:117] "RemoveContainer" containerID="f28a8e69ef57b9b7c639e4c1eaeaa76b528e468b65280cc01ce4c2dfd93d2b81" Nov 24 17:40:26 crc kubenswrapper[4808]: E1124 17:40:26.647906 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f28a8e69ef57b9b7c639e4c1eaeaa76b528e468b65280cc01ce4c2dfd93d2b81\": container with ID starting with f28a8e69ef57b9b7c639e4c1eaeaa76b528e468b65280cc01ce4c2dfd93d2b81 not found: ID does not exist" containerID="f28a8e69ef57b9b7c639e4c1eaeaa76b528e468b65280cc01ce4c2dfd93d2b81" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.647938 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f28a8e69ef57b9b7c639e4c1eaeaa76b528e468b65280cc01ce4c2dfd93d2b81"} err="failed to get container status \"f28a8e69ef57b9b7c639e4c1eaeaa76b528e468b65280cc01ce4c2dfd93d2b81\": rpc error: code = NotFound desc = could not find container \"f28a8e69ef57b9b7c639e4c1eaeaa76b528e468b65280cc01ce4c2dfd93d2b81\": container with ID starting with f28a8e69ef57b9b7c639e4c1eaeaa76b528e468b65280cc01ce4c2dfd93d2b81 not found: ID does not exist" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.647954 4808 scope.go:117] "RemoveContainer" containerID="20594e11b9fae81a8a39584f84ec9aa4b83bb6930e037c580a93728b43f8e251" Nov 24 17:40:26 crc kubenswrapper[4808]: E1124 17:40:26.648497 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20594e11b9fae81a8a39584f84ec9aa4b83bb6930e037c580a93728b43f8e251\": container with ID starting with 20594e11b9fae81a8a39584f84ec9aa4b83bb6930e037c580a93728b43f8e251 not found: ID does not exist" containerID="20594e11b9fae81a8a39584f84ec9aa4b83bb6930e037c580a93728b43f8e251" Nov 24 17:40:26 crc kubenswrapper[4808]: I1124 17:40:26.648530 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20594e11b9fae81a8a39584f84ec9aa4b83bb6930e037c580a93728b43f8e251"} err="failed to get container status \"20594e11b9fae81a8a39584f84ec9aa4b83bb6930e037c580a93728b43f8e251\": rpc error: code = NotFound desc = could not find container \"20594e11b9fae81a8a39584f84ec9aa4b83bb6930e037c580a93728b43f8e251\": container with ID starting with 20594e11b9fae81a8a39584f84ec9aa4b83bb6930e037c580a93728b43f8e251 not found: ID does not exist" Nov 24 17:40:28 crc kubenswrapper[4808]: I1124 17:40:28.356575 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2" path="/var/lib/kubelet/pods/8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2/volumes" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.226419 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-bbdb9b8fd-hfwr9"] Nov 24 17:40:36 crc kubenswrapper[4808]: E1124 17:40:36.227723 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15731e66-4629-4f34-af6b-83ae11044e5f" containerName="extract-utilities" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.227742 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="15731e66-4629-4f34-af6b-83ae11044e5f" containerName="extract-utilities" Nov 24 17:40:36 crc kubenswrapper[4808]: E1124 17:40:36.227757 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2" containerName="extract-content" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.227765 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2" containerName="extract-content" Nov 24 17:40:36 crc kubenswrapper[4808]: E1124 17:40:36.227779 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15731e66-4629-4f34-af6b-83ae11044e5f" containerName="registry-server" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.227787 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="15731e66-4629-4f34-af6b-83ae11044e5f" containerName="registry-server" Nov 24 17:40:36 crc kubenswrapper[4808]: E1124 17:40:36.227800 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a" containerName="extract-content" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.227808 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a" containerName="extract-content" Nov 24 17:40:36 crc kubenswrapper[4808]: E1124 17:40:36.227821 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2" containerName="extract-utilities" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.227830 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2" containerName="extract-utilities" Nov 24 17:40:36 crc kubenswrapper[4808]: E1124 17:40:36.227841 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2" containerName="registry-server" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.227848 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2" containerName="registry-server" Nov 24 17:40:36 crc kubenswrapper[4808]: E1124 17:40:36.227865 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a" containerName="extract-utilities" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.227872 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a" containerName="extract-utilities" Nov 24 17:40:36 crc kubenswrapper[4808]: E1124 17:40:36.227880 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15731e66-4629-4f34-af6b-83ae11044e5f" containerName="extract-content" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.227890 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="15731e66-4629-4f34-af6b-83ae11044e5f" containerName="extract-content" Nov 24 17:40:36 crc kubenswrapper[4808]: E1124 17:40:36.227902 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a" containerName="registry-server" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.227909 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a" containerName="registry-server" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.228081 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dfe5bb8-3dc5-4ddb-84fd-c078a3fd2b0a" containerName="registry-server" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.228094 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="15731e66-4629-4f34-af6b-83ae11044e5f" containerName="registry-server" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.228110 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e8a57e0-ef3f-4f8f-ae13-70b05ca9f9c2" containerName="registry-server" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.229085 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-bbdb9b8fd-hfwr9" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.234498 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-5qr7p" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.240238 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-942s7"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.242132 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-942s7" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.245228 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-bbdb9b8fd-hfwr9"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.252738 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-gdlq2" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.271771 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-942s7"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.287526 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-nv6j9"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.288825 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-nv6j9" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.293967 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-drdg2" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.311699 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-nv6j9"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.331810 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7fd7485997-ssvg7"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.337634 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7fd7485997-ssvg7" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.346160 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-bbpns" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.365940 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-758c497787-nnb8j"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.368674 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-758c497787-nnb8j" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.373609 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-q55sr" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.376735 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7fd7485997-ssvg7"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.401495 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svjvp\" (UniqueName: \"kubernetes.io/projected/bf340213-3bb6-4fc3-b8ed-9728393e9a5a-kube-api-access-svjvp\") pod \"cinder-operator-controller-manager-79856dc55c-942s7\" (UID: \"bf340213-3bb6-4fc3-b8ed-9728393e9a5a\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-942s7" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.401623 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2wdq\" (UniqueName: \"kubernetes.io/projected/2ece0883-b2fd-4309-b4bd-7639d95445e2-kube-api-access-q2wdq\") pod \"barbican-operator-controller-manager-bbdb9b8fd-hfwr9\" (UID: \"2ece0883-b2fd-4309-b4bd-7639d95445e2\") " pod="openstack-operators/barbican-operator-controller-manager-bbdb9b8fd-hfwr9" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.401657 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpbkj\" (UniqueName: \"kubernetes.io/projected/edd12c74-ba4f-4cb4-8531-b98607fdab41-kube-api-access-cpbkj\") pod \"designate-operator-controller-manager-7d695c9b56-nv6j9\" (UID: \"edd12c74-ba4f-4cb4-8531-b98607fdab41\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-nv6j9" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.411976 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-758c497787-nnb8j"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.428801 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-2fzkr"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.429945 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2fzkr" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.436858 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-6vpbj" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.456090 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-299sb"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.457358 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.459682 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.459940 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-zxjzn" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.481420 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q5xs5"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.485810 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q5xs5" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.487615 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-299sb"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.490926 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-z6hqh" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.520862 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvfls\" (UniqueName: \"kubernetes.io/projected/352d8a11-a201-4bc3-8d10-8378a5e04bfb-kube-api-access-jvfls\") pod \"horizon-operator-controller-manager-68c9694994-2fzkr\" (UID: \"352d8a11-a201-4bc3-8d10-8378a5e04bfb\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2fzkr" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.521064 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svjvp\" (UniqueName: \"kubernetes.io/projected/bf340213-3bb6-4fc3-b8ed-9728393e9a5a-kube-api-access-svjvp\") pod \"cinder-operator-controller-manager-79856dc55c-942s7\" (UID: \"bf340213-3bb6-4fc3-b8ed-9728393e9a5a\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-942s7" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.521158 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdb47\" (UniqueName: \"kubernetes.io/projected/55559441-f0b0-4353-bbcb-55342cf6ef53-kube-api-access-jdb47\") pod \"heat-operator-controller-manager-758c497787-nnb8j\" (UID: \"55559441-f0b0-4353-bbcb-55342cf6ef53\") " pod="openstack-operators/heat-operator-controller-manager-758c497787-nnb8j" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.521328 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2wdq\" (UniqueName: \"kubernetes.io/projected/2ece0883-b2fd-4309-b4bd-7639d95445e2-kube-api-access-q2wdq\") pod \"barbican-operator-controller-manager-bbdb9b8fd-hfwr9\" (UID: \"2ece0883-b2fd-4309-b4bd-7639d95445e2\") " pod="openstack-operators/barbican-operator-controller-manager-bbdb9b8fd-hfwr9" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.521378 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpbkj\" (UniqueName: \"kubernetes.io/projected/edd12c74-ba4f-4cb4-8531-b98607fdab41-kube-api-access-cpbkj\") pod \"designate-operator-controller-manager-7d695c9b56-nv6j9\" (UID: \"edd12c74-ba4f-4cb4-8531-b98607fdab41\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-nv6j9" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.521422 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfl2t\" (UniqueName: \"kubernetes.io/projected/fa1e211d-28e9-4ea1-beb5-6e3d0f5b8b9a-kube-api-access-mfl2t\") pod \"glance-operator-controller-manager-7fd7485997-ssvg7\" (UID: \"fa1e211d-28e9-4ea1-beb5-6e3d0f5b8b9a\") " pod="openstack-operators/glance-operator-controller-manager-7fd7485997-ssvg7" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.596566 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.596637 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.596692 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.599360 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q5xs5"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.609205 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpbkj\" (UniqueName: \"kubernetes.io/projected/edd12c74-ba4f-4cb4-8531-b98607fdab41-kube-api-access-cpbkj\") pod \"designate-operator-controller-manager-7d695c9b56-nv6j9\" (UID: \"edd12c74-ba4f-4cb4-8531-b98607fdab41\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-nv6j9" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.611995 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"36f63fea078d49c365c55146d216a74b13586224c5d10cc6c372c7c5a5288461"} pod="openshift-machine-config-operator/machine-config-daemon-ht87w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.612161 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" containerID="cri-o://36f63fea078d49c365c55146d216a74b13586224c5d10cc6c372c7c5a5288461" gracePeriod=600 Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.616440 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-tg4qk"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.622683 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58b1610a-0122-447e-a0a0-c2cb927a571f-cert\") pod \"infra-operator-controller-manager-57548d458d-299sb\" (UID: \"58b1610a-0122-447e-a0a0-c2cb927a571f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.644286 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvfls\" (UniqueName: \"kubernetes.io/projected/352d8a11-a201-4bc3-8d10-8378a5e04bfb-kube-api-access-jvfls\") pod \"horizon-operator-controller-manager-68c9694994-2fzkr\" (UID: \"352d8a11-a201-4bc3-8d10-8378a5e04bfb\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2fzkr" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.644438 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vbc2\" (UniqueName: \"kubernetes.io/projected/58d593d8-712e-4eb5-8b1c-5df55db3f1d9-kube-api-access-6vbc2\") pod \"ironic-operator-controller-manager-5bfcdc958c-q5xs5\" (UID: \"58d593d8-712e-4eb5-8b1c-5df55db3f1d9\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q5xs5" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.644531 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2zhk\" (UniqueName: \"kubernetes.io/projected/58b1610a-0122-447e-a0a0-c2cb927a571f-kube-api-access-x2zhk\") pod \"infra-operator-controller-manager-57548d458d-299sb\" (UID: \"58b1610a-0122-447e-a0a0-c2cb927a571f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.644569 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdb47\" (UniqueName: \"kubernetes.io/projected/55559441-f0b0-4353-bbcb-55342cf6ef53-kube-api-access-jdb47\") pod \"heat-operator-controller-manager-758c497787-nnb8j\" (UID: \"55559441-f0b0-4353-bbcb-55342cf6ef53\") " pod="openstack-operators/heat-operator-controller-manager-758c497787-nnb8j" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.644791 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfl2t\" (UniqueName: \"kubernetes.io/projected/fa1e211d-28e9-4ea1-beb5-6e3d0f5b8b9a-kube-api-access-mfl2t\") pod \"glance-operator-controller-manager-7fd7485997-ssvg7\" (UID: \"fa1e211d-28e9-4ea1-beb5-6e3d0f5b8b9a\") " pod="openstack-operators/glance-operator-controller-manager-7fd7485997-ssvg7" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.622927 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tg4qk" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.623778 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2wdq\" (UniqueName: \"kubernetes.io/projected/2ece0883-b2fd-4309-b4bd-7639d95445e2-kube-api-access-q2wdq\") pod \"barbican-operator-controller-manager-bbdb9b8fd-hfwr9\" (UID: \"2ece0883-b2fd-4309-b4bd-7639d95445e2\") " pod="openstack-operators/barbican-operator-controller-manager-bbdb9b8fd-hfwr9" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.635647 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-nv6j9" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.640092 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-8659db65fc-r4hcb"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.651361 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-8659db65fc-r4hcb" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.657126 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5f598f884b-xqgvh"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.658860 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5f598f884b-xqgvh" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.661989 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-bmbdb" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.662242 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-2pcfz" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.665737 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svjvp\" (UniqueName: \"kubernetes.io/projected/bf340213-3bb6-4fc3-b8ed-9728393e9a5a-kube-api-access-svjvp\") pod \"cinder-operator-controller-manager-79856dc55c-942s7\" (UID: \"bf340213-3bb6-4fc3-b8ed-9728393e9a5a\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-942s7" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.667382 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-cjf6g" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.676900 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-8659db65fc-r4hcb"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.704062 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-2fzkr"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.724390 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-tg4qk"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.724592 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvfls\" (UniqueName: \"kubernetes.io/projected/352d8a11-a201-4bc3-8d10-8378a5e04bfb-kube-api-access-jvfls\") pod \"horizon-operator-controller-manager-68c9694994-2fzkr\" (UID: \"352d8a11-a201-4bc3-8d10-8378a5e04bfb\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2fzkr" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.726856 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdb47\" (UniqueName: \"kubernetes.io/projected/55559441-f0b0-4353-bbcb-55342cf6ef53-kube-api-access-jdb47\") pod \"heat-operator-controller-manager-758c497787-nnb8j\" (UID: \"55559441-f0b0-4353-bbcb-55342cf6ef53\") " pod="openstack-operators/heat-operator-controller-manager-758c497787-nnb8j" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.751431 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vbc2\" (UniqueName: \"kubernetes.io/projected/58d593d8-712e-4eb5-8b1c-5df55db3f1d9-kube-api-access-6vbc2\") pod \"ironic-operator-controller-manager-5bfcdc958c-q5xs5\" (UID: \"58d593d8-712e-4eb5-8b1c-5df55db3f1d9\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q5xs5" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.751953 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2zhk\" (UniqueName: \"kubernetes.io/projected/58b1610a-0122-447e-a0a0-c2cb927a571f-kube-api-access-x2zhk\") pod \"infra-operator-controller-manager-57548d458d-299sb\" (UID: \"58b1610a-0122-447e-a0a0-c2cb927a571f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.751996 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln9tv\" (UniqueName: \"kubernetes.io/projected/01054a01-64a1-4c81-b03f-5dfdf55ee4cf-kube-api-access-ln9tv\") pod \"manila-operator-controller-manager-8659db65fc-r4hcb\" (UID: \"01054a01-64a1-4c81-b03f-5dfdf55ee4cf\") " pod="openstack-operators/manila-operator-controller-manager-8659db65fc-r4hcb" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.752110 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58b1610a-0122-447e-a0a0-c2cb927a571f-cert\") pod \"infra-operator-controller-manager-57548d458d-299sb\" (UID: \"58b1610a-0122-447e-a0a0-c2cb927a571f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.752155 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm9kx\" (UniqueName: \"kubernetes.io/projected/9d1d2914-59b8-4660-80a4-d2a1cc17c2bc-kube-api-access-sm9kx\") pod \"keystone-operator-controller-manager-748dc6576f-tg4qk\" (UID: \"9d1d2914-59b8-4660-80a4-d2a1cc17c2bc\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tg4qk" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.752182 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gltn\" (UniqueName: \"kubernetes.io/projected/ac1adbdb-b43a-4d5a-b599-7e2993f72a86-kube-api-access-8gltn\") pod \"mariadb-operator-controller-manager-5f598f884b-xqgvh\" (UID: \"ac1adbdb-b43a-4d5a-b599-7e2993f72a86\") " pod="openstack-operators/mariadb-operator-controller-manager-5f598f884b-xqgvh" Nov 24 17:40:36 crc kubenswrapper[4808]: E1124 17:40:36.752691 4808 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 24 17:40:36 crc kubenswrapper[4808]: E1124 17:40:36.752741 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58b1610a-0122-447e-a0a0-c2cb927a571f-cert podName:58b1610a-0122-447e-a0a0-c2cb927a571f nodeName:}" failed. No retries permitted until 2025-11-24 17:40:37.252720531 +0000 UTC m=+829.850388343 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/58b1610a-0122-447e-a0a0-c2cb927a571f-cert") pod "infra-operator-controller-manager-57548d458d-299sb" (UID: "58b1610a-0122-447e-a0a0-c2cb927a571f") : secret "infra-operator-webhook-server-cert" not found Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.762266 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfl2t\" (UniqueName: \"kubernetes.io/projected/fa1e211d-28e9-4ea1-beb5-6e3d0f5b8b9a-kube-api-access-mfl2t\") pod \"glance-operator-controller-manager-7fd7485997-ssvg7\" (UID: \"fa1e211d-28e9-4ea1-beb5-6e3d0f5b8b9a\") " pod="openstack-operators/glance-operator-controller-manager-7fd7485997-ssvg7" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.814904 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2fzkr" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.826933 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5f598f884b-xqgvh"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.831771 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2zhk\" (UniqueName: \"kubernetes.io/projected/58b1610a-0122-447e-a0a0-c2cb927a571f-kube-api-access-x2zhk\") pod \"infra-operator-controller-manager-57548d458d-299sb\" (UID: \"58b1610a-0122-447e-a0a0-c2cb927a571f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.836665 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-qrr4f"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.838319 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-qrr4f" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.841097 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-rcwzz"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.860867 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-qrr4f"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.860929 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-rcwzz"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.861069 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-rcwzz" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.861270 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-b5vdf"] Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.861559 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-r4mg4" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.862655 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-bbdb9b8fd-hfwr9" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.862706 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-b5vdf" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.863573 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm9kx\" (UniqueName: \"kubernetes.io/projected/9d1d2914-59b8-4660-80a4-d2a1cc17c2bc-kube-api-access-sm9kx\") pod \"keystone-operator-controller-manager-748dc6576f-tg4qk\" (UID: \"9d1d2914-59b8-4660-80a4-d2a1cc17c2bc\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tg4qk" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.863606 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gltn\" (UniqueName: \"kubernetes.io/projected/ac1adbdb-b43a-4d5a-b599-7e2993f72a86-kube-api-access-8gltn\") pod \"mariadb-operator-controller-manager-5f598f884b-xqgvh\" (UID: \"ac1adbdb-b43a-4d5a-b599-7e2993f72a86\") " pod="openstack-operators/mariadb-operator-controller-manager-5f598f884b-xqgvh" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.863765 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln9tv\" (UniqueName: \"kubernetes.io/projected/01054a01-64a1-4c81-b03f-5dfdf55ee4cf-kube-api-access-ln9tv\") pod \"manila-operator-controller-manager-8659db65fc-r4hcb\" (UID: \"01054a01-64a1-4c81-b03f-5dfdf55ee4cf\") " pod="openstack-operators/manila-operator-controller-manager-8659db65fc-r4hcb" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.866490 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vbc2\" (UniqueName: \"kubernetes.io/projected/58d593d8-712e-4eb5-8b1c-5df55db3f1d9-kube-api-access-6vbc2\") pod \"ironic-operator-controller-manager-5bfcdc958c-q5xs5\" (UID: \"58d593d8-712e-4eb5-8b1c-5df55db3f1d9\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q5xs5" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.878683 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-942s7" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.886448 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-nsb88" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.886700 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-ft6fd" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.949300 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm9kx\" (UniqueName: \"kubernetes.io/projected/9d1d2914-59b8-4660-80a4-d2a1cc17c2bc-kube-api-access-sm9kx\") pod \"keystone-operator-controller-manager-748dc6576f-tg4qk\" (UID: \"9d1d2914-59b8-4660-80a4-d2a1cc17c2bc\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tg4qk" Nov 24 17:40:36 crc kubenswrapper[4808]: I1124 17:40:36.971390 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7fd7485997-ssvg7" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.000648 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln9tv\" (UniqueName: \"kubernetes.io/projected/01054a01-64a1-4c81-b03f-5dfdf55ee4cf-kube-api-access-ln9tv\") pod \"manila-operator-controller-manager-8659db65fc-r4hcb\" (UID: \"01054a01-64a1-4c81-b03f-5dfdf55ee4cf\") " pod="openstack-operators/manila-operator-controller-manager-8659db65fc-r4hcb" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.001143 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-b5vdf"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.001185 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.002451 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.003650 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhbls\" (UniqueName: \"kubernetes.io/projected/8a4f09fc-1405-4a74-be87-1b94ee06742a-kube-api-access-fhbls\") pod \"neutron-operator-controller-manager-6fdcddb789-qrr4f\" (UID: \"8a4f09fc-1405-4a74-be87-1b94ee06742a\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-qrr4f" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.003855 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2gmf\" (UniqueName: \"kubernetes.io/projected/47c249db-b332-44cf-baa5-811e04500c0f-kube-api-access-h2gmf\") pod \"nova-operator-controller-manager-79556f57fc-rcwzz\" (UID: \"47c249db-b332-44cf-baa5-811e04500c0f\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-rcwzz" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.003956 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6wzr\" (UniqueName: \"kubernetes.io/projected/af79a806-a6fb-40a1-8ac9-3fccd532c1fc-kube-api-access-x6wzr\") pod \"octavia-operator-controller-manager-fd75fd47d-b5vdf\" (UID: \"af79a806-a6fb-40a1-8ac9-3fccd532c1fc\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-b5vdf" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.007159 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gltn\" (UniqueName: \"kubernetes.io/projected/ac1adbdb-b43a-4d5a-b599-7e2993f72a86-kube-api-access-8gltn\") pod \"mariadb-operator-controller-manager-5f598f884b-xqgvh\" (UID: \"ac1adbdb-b43a-4d5a-b599-7e2993f72a86\") " pod="openstack-operators/mariadb-operator-controller-manager-5f598f884b-xqgvh" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.018495 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-758c497787-nnb8j" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.019357 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.019456 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-mxxj6" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.023060 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-668bd84cc7-lm5c4"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.029251 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-668bd84cc7-lm5c4" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.049784 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.052856 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.060097 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.067627 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-wq75p" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.067821 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-hthrk" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.074525 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-668bd84cc7-lm5c4"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.114961 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28sgn\" (UniqueName: \"kubernetes.io/projected/dc05be1a-a87d-4829-b2a8-c9d454b7d343-kube-api-access-28sgn\") pod \"ovn-operator-controller-manager-668bd84cc7-lm5c4\" (UID: \"dc05be1a-a87d-4829-b2a8-c9d454b7d343\") " pod="openstack-operators/ovn-operator-controller-manager-668bd84cc7-lm5c4" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.115558 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-856jn\" (UniqueName: \"kubernetes.io/projected/c796b113-a891-48b2-9a11-50a2577baf7e-kube-api-access-856jn\") pod \"openstack-baremetal-operator-controller-manager-7d654b7b69f77rn\" (UID: \"c796b113-a891-48b2-9a11-50a2577baf7e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.115668 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2gmf\" (UniqueName: \"kubernetes.io/projected/47c249db-b332-44cf-baa5-811e04500c0f-kube-api-access-h2gmf\") pod \"nova-operator-controller-manager-79556f57fc-rcwzz\" (UID: \"47c249db-b332-44cf-baa5-811e04500c0f\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-rcwzz" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.115799 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6wzr\" (UniqueName: \"kubernetes.io/projected/af79a806-a6fb-40a1-8ac9-3fccd532c1fc-kube-api-access-x6wzr\") pod \"octavia-operator-controller-manager-fd75fd47d-b5vdf\" (UID: \"af79a806-a6fb-40a1-8ac9-3fccd532c1fc\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-b5vdf" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.115858 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhbls\" (UniqueName: \"kubernetes.io/projected/8a4f09fc-1405-4a74-be87-1b94ee06742a-kube-api-access-fhbls\") pod \"neutron-operator-controller-manager-6fdcddb789-qrr4f\" (UID: \"8a4f09fc-1405-4a74-be87-1b94ee06742a\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-qrr4f" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.115882 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c796b113-a891-48b2-9a11-50a2577baf7e-cert\") pod \"openstack-baremetal-operator-controller-manager-7d654b7b69f77rn\" (UID: \"c796b113-a891-48b2-9a11-50a2577baf7e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.115967 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbsvw\" (UniqueName: \"kubernetes.io/projected/321df5bf-732b-4982-841b-3a9a0ddb73f2-kube-api-access-bbsvw\") pod \"placement-operator-controller-manager-5db546f9d9-zgnpn\" (UID: \"321df5bf-732b-4982-841b-3a9a0ddb73f2\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.138619 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mdswb"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.139755 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tg4qk" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.139972 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mdswb" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.140459 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q5xs5" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.161816 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhbls\" (UniqueName: \"kubernetes.io/projected/8a4f09fc-1405-4a74-be87-1b94ee06742a-kube-api-access-fhbls\") pod \"neutron-operator-controller-manager-6fdcddb789-qrr4f\" (UID: \"8a4f09fc-1405-4a74-be87-1b94ee06742a\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-qrr4f" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.164276 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-627xm" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.177679 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6wzr\" (UniqueName: \"kubernetes.io/projected/af79a806-a6fb-40a1-8ac9-3fccd532c1fc-kube-api-access-x6wzr\") pod \"octavia-operator-controller-manager-fd75fd47d-b5vdf\" (UID: \"af79a806-a6fb-40a1-8ac9-3fccd532c1fc\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-b5vdf" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.177758 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.190518 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-8659db65fc-r4hcb" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.214408 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2gmf\" (UniqueName: \"kubernetes.io/projected/47c249db-b332-44cf-baa5-811e04500c0f-kube-api-access-h2gmf\") pod \"nova-operator-controller-manager-79556f57fc-rcwzz\" (UID: \"47c249db-b332-44cf-baa5-811e04500c0f\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-rcwzz" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.216817 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.217784 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-856jn\" (UniqueName: \"kubernetes.io/projected/c796b113-a891-48b2-9a11-50a2577baf7e-kube-api-access-856jn\") pod \"openstack-baremetal-operator-controller-manager-7d654b7b69f77rn\" (UID: \"c796b113-a891-48b2-9a11-50a2577baf7e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.217877 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c796b113-a891-48b2-9a11-50a2577baf7e-cert\") pod \"openstack-baremetal-operator-controller-manager-7d654b7b69f77rn\" (UID: \"c796b113-a891-48b2-9a11-50a2577baf7e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.217912 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbsvw\" (UniqueName: \"kubernetes.io/projected/321df5bf-732b-4982-841b-3a9a0ddb73f2-kube-api-access-bbsvw\") pod \"placement-operator-controller-manager-5db546f9d9-zgnpn\" (UID: \"321df5bf-732b-4982-841b-3a9a0ddb73f2\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.217960 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28sgn\" (UniqueName: \"kubernetes.io/projected/dc05be1a-a87d-4829-b2a8-c9d454b7d343-kube-api-access-28sgn\") pod \"ovn-operator-controller-manager-668bd84cc7-lm5c4\" (UID: \"dc05be1a-a87d-4829-b2a8-c9d454b7d343\") " pod="openstack-operators/ovn-operator-controller-manager-668bd84cc7-lm5c4" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.217981 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfq8w\" (UniqueName: \"kubernetes.io/projected/2c56d9f3-3ab5-4f3c-b250-e926807dbfc3-kube-api-access-hfq8w\") pod \"swift-operator-controller-manager-6fdc4fcf86-mdswb\" (UID: \"2c56d9f3-3ab5-4f3c-b250-e926807dbfc3\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mdswb" Nov 24 17:40:37 crc kubenswrapper[4808]: E1124 17:40:37.218431 4808 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 17:40:37 crc kubenswrapper[4808]: E1124 17:40:37.218481 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c796b113-a891-48b2-9a11-50a2577baf7e-cert podName:c796b113-a891-48b2-9a11-50a2577baf7e nodeName:}" failed. No retries permitted until 2025-11-24 17:40:37.718467511 +0000 UTC m=+830.316135313 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c796b113-a891-48b2-9a11-50a2577baf7e-cert") pod "openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" (UID: "c796b113-a891-48b2-9a11-50a2577baf7e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.219054 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.222740 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-b5vdf" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.224791 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-ffq8r" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.256311 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mdswb"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.258475 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5f598f884b-xqgvh" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.290874 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-856jn\" (UniqueName: \"kubernetes.io/projected/c796b113-a891-48b2-9a11-50a2577baf7e-kube-api-access-856jn\") pod \"openstack-baremetal-operator-controller-manager-7d654b7b69f77rn\" (UID: \"c796b113-a891-48b2-9a11-50a2577baf7e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.294975 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbsvw\" (UniqueName: \"kubernetes.io/projected/321df5bf-732b-4982-841b-3a9a0ddb73f2-kube-api-access-bbsvw\") pod \"placement-operator-controller-manager-5db546f9d9-zgnpn\" (UID: \"321df5bf-732b-4982-841b-3a9a0ddb73f2\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.300319 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28sgn\" (UniqueName: \"kubernetes.io/projected/dc05be1a-a87d-4829-b2a8-c9d454b7d343-kube-api-access-28sgn\") pod \"ovn-operator-controller-manager-668bd84cc7-lm5c4\" (UID: \"dc05be1a-a87d-4829-b2a8-c9d454b7d343\") " pod="openstack-operators/ovn-operator-controller-manager-668bd84cc7-lm5c4" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.309601 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.331171 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tjkp\" (UniqueName: \"kubernetes.io/projected/ec05f4fb-ca34-4f68-b94e-95b08b21d57b-kube-api-access-7tjkp\") pod \"telemetry-operator-controller-manager-567f98c9d-9wwt6\" (UID: \"ec05f4fb-ca34-4f68-b94e-95b08b21d57b\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.331341 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58b1610a-0122-447e-a0a0-c2cb927a571f-cert\") pod \"infra-operator-controller-manager-57548d458d-299sb\" (UID: \"58b1610a-0122-447e-a0a0-c2cb927a571f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.331438 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfq8w\" (UniqueName: \"kubernetes.io/projected/2c56d9f3-3ab5-4f3c-b250-e926807dbfc3-kube-api-access-hfq8w\") pod \"swift-operator-controller-manager-6fdc4fcf86-mdswb\" (UID: \"2c56d9f3-3ab5-4f3c-b250-e926807dbfc3\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mdswb" Nov 24 17:40:37 crc kubenswrapper[4808]: E1124 17:40:37.332424 4808 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 24 17:40:37 crc kubenswrapper[4808]: E1124 17:40:37.332494 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58b1610a-0122-447e-a0a0-c2cb927a571f-cert podName:58b1610a-0122-447e-a0a0-c2cb927a571f nodeName:}" failed. No retries permitted until 2025-11-24 17:40:38.332469642 +0000 UTC m=+830.930137444 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/58b1610a-0122-447e-a0a0-c2cb927a571f-cert") pod "infra-operator-controller-manager-57548d458d-299sb" (UID: "58b1610a-0122-447e-a0a0-c2cb927a571f") : secret "infra-operator-webhook-server-cert" not found Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.356060 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-66sz4"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.357616 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-qrr4f" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.387814 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-66sz4" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.397771 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-hvflw" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.450820 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-668bd84cc7-lm5c4" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.463324 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfq8w\" (UniqueName: \"kubernetes.io/projected/2c56d9f3-3ab5-4f3c-b250-e926807dbfc3-kube-api-access-hfq8w\") pod \"swift-operator-controller-manager-6fdc4fcf86-mdswb\" (UID: \"2c56d9f3-3ab5-4f3c-b250-e926807dbfc3\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mdswb" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.483635 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-rcwzz" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.522458 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-66sz4"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.556004 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tjkp\" (UniqueName: \"kubernetes.io/projected/ec05f4fb-ca34-4f68-b94e-95b08b21d57b-kube-api-access-7tjkp\") pod \"telemetry-operator-controller-manager-567f98c9d-9wwt6\" (UID: \"ec05f4fb-ca34-4f68-b94e-95b08b21d57b\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.573161 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.612334 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.616972 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tjkp\" (UniqueName: \"kubernetes.io/projected/ec05f4fb-ca34-4f68-b94e-95b08b21d57b-kube-api-access-7tjkp\") pod \"telemetry-operator-controller-manager-567f98c9d-9wwt6\" (UID: \"ec05f4fb-ca34-4f68-b94e-95b08b21d57b\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.618241 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.623717 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-2mjt2" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.623862 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.697271 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2tcp\" (UniqueName: \"kubernetes.io/projected/eb8cf6b6-33aa-427e-9d07-9f95c58c7d30-kube-api-access-m2tcp\") pod \"test-operator-controller-manager-5cd6c7f4c8-66sz4\" (UID: \"eb8cf6b6-33aa-427e-9d07-9f95c58c7d30\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-66sz4" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.697330 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhbmj\" (UniqueName: \"kubernetes.io/projected/c62f31fd-04d1-48cc-ae7a-98e1e64c5af7-kube-api-access-qhbmj\") pod \"watcher-operator-controller-manager-656dcb59d4-wknfh\" (UID: \"c62f31fd-04d1-48cc-ae7a-98e1e64c5af7\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.762253 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.763632 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.769752 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-77bmj" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.770056 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.770159 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.791168 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerID="36f63fea078d49c365c55146d216a74b13586224c5d10cc6c372c7c5a5288461" exitCode=0 Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.791221 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerDied","Data":"36f63fea078d49c365c55146d216a74b13586224c5d10cc6c372c7c5a5288461"} Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.791250 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"216625ce434c33017d87034c7bea8e120016af1d6be08176847569ea1821e3e1"} Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.791269 4808 scope.go:117] "RemoveContainer" containerID="8b5c11a71abbce839aaa56f6f8510529f68981025ad2c3fda226177d751485f2" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.796064 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.799555 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2tcp\" (UniqueName: \"kubernetes.io/projected/eb8cf6b6-33aa-427e-9d07-9f95c58c7d30-kube-api-access-m2tcp\") pod \"test-operator-controller-manager-5cd6c7f4c8-66sz4\" (UID: \"eb8cf6b6-33aa-427e-9d07-9f95c58c7d30\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-66sz4" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.799614 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhbmj\" (UniqueName: \"kubernetes.io/projected/c62f31fd-04d1-48cc-ae7a-98e1e64c5af7-kube-api-access-qhbmj\") pod \"watcher-operator-controller-manager-656dcb59d4-wknfh\" (UID: \"c62f31fd-04d1-48cc-ae7a-98e1e64c5af7\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.799665 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c796b113-a891-48b2-9a11-50a2577baf7e-cert\") pod \"openstack-baremetal-operator-controller-manager-7d654b7b69f77rn\" (UID: \"c796b113-a891-48b2-9a11-50a2577baf7e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" Nov 24 17:40:37 crc kubenswrapper[4808]: E1124 17:40:37.799859 4808 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 17:40:37 crc kubenswrapper[4808]: E1124 17:40:37.799919 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c796b113-a891-48b2-9a11-50a2577baf7e-cert podName:c796b113-a891-48b2-9a11-50a2577baf7e nodeName:}" failed. No retries permitted until 2025-11-24 17:40:38.799899228 +0000 UTC m=+831.397567030 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c796b113-a891-48b2-9a11-50a2577baf7e-cert") pod "openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" (UID: "c796b113-a891-48b2-9a11-50a2577baf7e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.827669 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4r97r"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.829406 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4r97r" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.834619 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-x72pd" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.840895 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhbmj\" (UniqueName: \"kubernetes.io/projected/c62f31fd-04d1-48cc-ae7a-98e1e64c5af7-kube-api-access-qhbmj\") pod \"watcher-operator-controller-manager-656dcb59d4-wknfh\" (UID: \"c62f31fd-04d1-48cc-ae7a-98e1e64c5af7\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.841000 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4r97r"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.841907 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2tcp\" (UniqueName: \"kubernetes.io/projected/eb8cf6b6-33aa-427e-9d07-9f95c58c7d30-kube-api-access-m2tcp\") pod \"test-operator-controller-manager-5cd6c7f4c8-66sz4\" (UID: \"eb8cf6b6-33aa-427e-9d07-9f95c58c7d30\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-66sz4" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.856825 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-nv6j9"] Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.900863 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmf64\" (UniqueName: \"kubernetes.io/projected/e597e7fc-3b56-493b-bcda-10669936636a-kube-api-access-gmf64\") pod \"openstack-operator-controller-manager-79c6ffd6b7-vvfcw\" (UID: \"e597e7fc-3b56-493b-bcda-10669936636a\") " pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.900938 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-metrics-certs\") pod \"openstack-operator-controller-manager-79c6ffd6b7-vvfcw\" (UID: \"e597e7fc-3b56-493b-bcda-10669936636a\") " pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.901094 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-webhook-certs\") pod \"openstack-operator-controller-manager-79c6ffd6b7-vvfcw\" (UID: \"e597e7fc-3b56-493b-bcda-10669936636a\") " pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.924084 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-66sz4" Nov 24 17:40:37 crc kubenswrapper[4808]: I1124 17:40:37.993735 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mdswb" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.004516 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-webhook-certs\") pod \"openstack-operator-controller-manager-79c6ffd6b7-vvfcw\" (UID: \"e597e7fc-3b56-493b-bcda-10669936636a\") " pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.004719 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmf64\" (UniqueName: \"kubernetes.io/projected/e597e7fc-3b56-493b-bcda-10669936636a-kube-api-access-gmf64\") pod \"openstack-operator-controller-manager-79c6ffd6b7-vvfcw\" (UID: \"e597e7fc-3b56-493b-bcda-10669936636a\") " pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.004751 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-metrics-certs\") pod \"openstack-operator-controller-manager-79c6ffd6b7-vvfcw\" (UID: \"e597e7fc-3b56-493b-bcda-10669936636a\") " pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.004866 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcx5v\" (UniqueName: \"kubernetes.io/projected/5d6ac280-f3af-4ee8-8933-928ed0c501fc-kube-api-access-bcx5v\") pod \"rabbitmq-cluster-operator-manager-668c99d594-4r97r\" (UID: \"5d6ac280-f3af-4ee8-8933-928ed0c501fc\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4r97r" Nov 24 17:40:38 crc kubenswrapper[4808]: E1124 17:40:38.006392 4808 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 24 17:40:38 crc kubenswrapper[4808]: E1124 17:40:38.006443 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-webhook-certs podName:e597e7fc-3b56-493b-bcda-10669936636a nodeName:}" failed. No retries permitted until 2025-11-24 17:40:38.50642795 +0000 UTC m=+831.104095752 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-webhook-certs") pod "openstack-operator-controller-manager-79c6ffd6b7-vvfcw" (UID: "e597e7fc-3b56-493b-bcda-10669936636a") : secret "webhook-server-cert" not found Nov 24 17:40:38 crc kubenswrapper[4808]: E1124 17:40:38.007531 4808 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 24 17:40:38 crc kubenswrapper[4808]: E1124 17:40:38.007579 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-metrics-certs podName:e597e7fc-3b56-493b-bcda-10669936636a nodeName:}" failed. No retries permitted until 2025-11-24 17:40:38.507571451 +0000 UTC m=+831.105239253 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-metrics-certs") pod "openstack-operator-controller-manager-79c6ffd6b7-vvfcw" (UID: "e597e7fc-3b56-493b-bcda-10669936636a") : secret "metrics-server-cert" not found Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.032821 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.048798 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmf64\" (UniqueName: \"kubernetes.io/projected/e597e7fc-3b56-493b-bcda-10669936636a-kube-api-access-gmf64\") pod \"openstack-operator-controller-manager-79c6ffd6b7-vvfcw\" (UID: \"e597e7fc-3b56-493b-bcda-10669936636a\") " pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.107116 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcx5v\" (UniqueName: \"kubernetes.io/projected/5d6ac280-f3af-4ee8-8933-928ed0c501fc-kube-api-access-bcx5v\") pod \"rabbitmq-cluster-operator-manager-668c99d594-4r97r\" (UID: \"5d6ac280-f3af-4ee8-8933-928ed0c501fc\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4r97r" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.142129 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcx5v\" (UniqueName: \"kubernetes.io/projected/5d6ac280-f3af-4ee8-8933-928ed0c501fc-kube-api-access-bcx5v\") pod \"rabbitmq-cluster-operator-manager-668c99d594-4r97r\" (UID: \"5d6ac280-f3af-4ee8-8933-928ed0c501fc\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4r97r" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.228684 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.246834 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4r97r" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.413622 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58b1610a-0122-447e-a0a0-c2cb927a571f-cert\") pod \"infra-operator-controller-manager-57548d458d-299sb\" (UID: \"58b1610a-0122-447e-a0a0-c2cb927a571f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.423224 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58b1610a-0122-447e-a0a0-c2cb927a571f-cert\") pod \"infra-operator-controller-manager-57548d458d-299sb\" (UID: \"58b1610a-0122-447e-a0a0-c2cb927a571f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.454203 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-2fzkr"] Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.465927 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7fd7485997-ssvg7"] Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.487137 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-bbdb9b8fd-hfwr9"] Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.519696 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-metrics-certs\") pod \"openstack-operator-controller-manager-79c6ffd6b7-vvfcw\" (UID: \"e597e7fc-3b56-493b-bcda-10669936636a\") " pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.519847 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-webhook-certs\") pod \"openstack-operator-controller-manager-79c6ffd6b7-vvfcw\" (UID: \"e597e7fc-3b56-493b-bcda-10669936636a\") " pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:38 crc kubenswrapper[4808]: E1124 17:40:38.520187 4808 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 24 17:40:38 crc kubenswrapper[4808]: E1124 17:40:38.520232 4808 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 24 17:40:38 crc kubenswrapper[4808]: E1124 17:40:38.520285 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-webhook-certs podName:e597e7fc-3b56-493b-bcda-10669936636a nodeName:}" failed. No retries permitted until 2025-11-24 17:40:39.520256181 +0000 UTC m=+832.117923983 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-webhook-certs") pod "openstack-operator-controller-manager-79c6ffd6b7-vvfcw" (UID: "e597e7fc-3b56-493b-bcda-10669936636a") : secret "webhook-server-cert" not found Nov 24 17:40:38 crc kubenswrapper[4808]: E1124 17:40:38.520315 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-metrics-certs podName:e597e7fc-3b56-493b-bcda-10669936636a nodeName:}" failed. No retries permitted until 2025-11-24 17:40:39.520294752 +0000 UTC m=+832.117962554 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-metrics-certs") pod "openstack-operator-controller-manager-79c6ffd6b7-vvfcw" (UID: "e597e7fc-3b56-493b-bcda-10669936636a") : secret "metrics-server-cert" not found Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.594584 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.675910 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-758c497787-nnb8j"] Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.779682 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-942s7"] Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.834591 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c796b113-a891-48b2-9a11-50a2577baf7e-cert\") pod \"openstack-baremetal-operator-controller-manager-7d654b7b69f77rn\" (UID: \"c796b113-a891-48b2-9a11-50a2577baf7e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.844725 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c796b113-a891-48b2-9a11-50a2577baf7e-cert\") pod \"openstack-baremetal-operator-controller-manager-7d654b7b69f77rn\" (UID: \"c796b113-a891-48b2-9a11-50a2577baf7e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.876469 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-758c497787-nnb8j" event={"ID":"55559441-f0b0-4353-bbcb-55342cf6ef53","Type":"ContainerStarted","Data":"01c6a5024a95b79613deba2b390509e3509852d4223eae4b7f56c972c7eb02f1"} Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.881854 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7fd7485997-ssvg7" event={"ID":"fa1e211d-28e9-4ea1-beb5-6e3d0f5b8b9a","Type":"ContainerStarted","Data":"cba85b78d20787ea9deae43ea8e987267df471c09af922f1b26794b2512cd3e6"} Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.892583 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-bbdb9b8fd-hfwr9" event={"ID":"2ece0883-b2fd-4309-b4bd-7639d95445e2","Type":"ContainerStarted","Data":"38c0492a3c51a679e73889ddf6ebf1e6b08ecdcddf4216fc978478597bb83ea5"} Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.913494 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-nv6j9" event={"ID":"edd12c74-ba4f-4cb4-8531-b98607fdab41","Type":"ContainerStarted","Data":"467013cc137476335591371811e1ced03cb8c43ad32ead4df1ec09814e433c73"} Nov 24 17:40:38 crc kubenswrapper[4808]: I1124 17:40:38.915946 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2fzkr" event={"ID":"352d8a11-a201-4bc3-8d10-8378a5e04bfb","Type":"ContainerStarted","Data":"8e68a37801a2a7840c1fa7b9f038dd4b2af938eea1c5b089a878e03666af3052"} Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.016810 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-8659db65fc-r4hcb"] Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.077384 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-b5vdf"] Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.098412 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-tg4qk"] Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.106712 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-qrr4f"] Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.119586 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.122626 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q5xs5"] Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.149737 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5f598f884b-xqgvh"] Nov 24 17:40:39 crc kubenswrapper[4808]: W1124 17:40:39.187739 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac1adbdb_b43a_4d5a_b599_7e2993f72a86.slice/crio-82bf2b352e7984d0e3c23e729aa01a6f521cd2ff13f5531367d9462d5e493d54 WatchSource:0}: Error finding container 82bf2b352e7984d0e3c23e729aa01a6f521cd2ff13f5531367d9462d5e493d54: Status 404 returned error can't find the container with id 82bf2b352e7984d0e3c23e729aa01a6f521cd2ff13f5531367d9462d5e493d54 Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.206388 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-668bd84cc7-lm5c4"] Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.286566 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn"] Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.296161 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-66sz4"] Nov 24 17:40:39 crc kubenswrapper[4808]: W1124 17:40:39.297159 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb8cf6b6_33aa_427e_9d07_9f95c58c7d30.slice/crio-a0397bf8bd8ac515387bf46afaf0f987622910a14e22dba0943843b01b48d393 WatchSource:0}: Error finding container a0397bf8bd8ac515387bf46afaf0f987622910a14e22dba0943843b01b48d393: Status 404 returned error can't find the container with id a0397bf8bd8ac515387bf46afaf0f987622910a14e22dba0943843b01b48d393 Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.302484 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-rcwzz"] Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.325523 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bbsvw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5db546f9d9-zgnpn_openstack-operators(321df5bf-732b-4982-841b-3a9a0ddb73f2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.327561 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bbsvw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5db546f9d9-zgnpn_openstack-operators(321df5bf-732b-4982-841b-3a9a0ddb73f2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.328682 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn" podUID="321df5bf-732b-4982-841b-3a9a0ddb73f2" Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.393454 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6"] Nov 24 17:40:39 crc kubenswrapper[4808]: W1124 17:40:39.396618 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec05f4fb_ca34_4f68_b94e_95b08b21d57b.slice/crio-ecb1220c0dabcc1b66c12a5ac29cdc2d658c4f6d7e4e3b13d7d8ffefbac7d116 WatchSource:0}: Error finding container ecb1220c0dabcc1b66c12a5ac29cdc2d658c4f6d7e4e3b13d7d8ffefbac7d116: Status 404 returned error can't find the container with id ecb1220c0dabcc1b66c12a5ac29cdc2d658c4f6d7e4e3b13d7d8ffefbac7d116 Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.425322 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh"] Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.425847 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7tjkp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-567f98c9d-9wwt6_openstack-operators(ec05f4fb-ca34-4f68-b94e-95b08b21d57b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.443336 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7tjkp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-567f98c9d-9wwt6_openstack-operators(ec05f4fb-ca34-4f68-b94e-95b08b21d57b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.444736 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6" podUID="ec05f4fb-ca34-4f68-b94e-95b08b21d57b" Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.447199 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mdswb"] Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.456092 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4r97r"] Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.468978 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-299sb"] Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.474746 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qhbmj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-656dcb59d4-wknfh_openstack-operators(c62f31fd-04d1-48cc-ae7a-98e1e64c5af7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.481086 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x2zhk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-57548d458d-299sb_openstack-operators(58b1610a-0122-447e-a0a0-c2cb927a571f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.481204 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qhbmj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-656dcb59d4-wknfh_openstack-operators(c62f31fd-04d1-48cc-ae7a-98e1e64c5af7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.481295 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hfq8w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6fdc4fcf86-mdswb_openstack-operators(2c56d9f3-3ab5-4f3c-b250-e926807dbfc3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.482396 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh" podUID="c62f31fd-04d1-48cc-ae7a-98e1e64c5af7" Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.483116 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x2zhk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-57548d458d-299sb_openstack-operators(58b1610a-0122-447e-a0a0-c2cb927a571f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.484295 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" podUID="58b1610a-0122-447e-a0a0-c2cb927a571f" Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.490333 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bcx5v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-4r97r_openstack-operators(5d6ac280-f3af-4ee8-8933-928ed0c501fc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.491893 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4r97r" podUID="5d6ac280-f3af-4ee8-8933-928ed0c501fc" Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.589901 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-webhook-certs\") pod \"openstack-operator-controller-manager-79c6ffd6b7-vvfcw\" (UID: \"e597e7fc-3b56-493b-bcda-10669936636a\") " pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.589965 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-metrics-certs\") pod \"openstack-operator-controller-manager-79c6ffd6b7-vvfcw\" (UID: \"e597e7fc-3b56-493b-bcda-10669936636a\") " pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.590091 4808 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.590137 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-metrics-certs podName:e597e7fc-3b56-493b-bcda-10669936636a nodeName:}" failed. No retries permitted until 2025-11-24 17:40:41.590123981 +0000 UTC m=+834.187791783 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-metrics-certs") pod "openstack-operator-controller-manager-79c6ffd6b7-vvfcw" (UID: "e597e7fc-3b56-493b-bcda-10669936636a") : secret "metrics-server-cert" not found Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.608619 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-webhook-certs\") pod \"openstack-operator-controller-manager-79c6ffd6b7-vvfcw\" (UID: \"e597e7fc-3b56-493b-bcda-10669936636a\") " pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.868870 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn"] Nov 24 17:40:39 crc kubenswrapper[4808]: W1124 17:40:39.924987 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc796b113_a891_48b2_9a11_50a2577baf7e.slice/crio-d589ea2d06cab1c7a7ed1605d0e61baadbc0abc8f9da58ab59b372123a764b2d WatchSource:0}: Error finding container d589ea2d06cab1c7a7ed1605d0e61baadbc0abc8f9da58ab59b372123a764b2d: Status 404 returned error can't find the container with id d589ea2d06cab1c7a7ed1605d0e61baadbc0abc8f9da58ab59b372123a764b2d Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.963366 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5f598f884b-xqgvh" event={"ID":"ac1adbdb-b43a-4d5a-b599-7e2993f72a86","Type":"ContainerStarted","Data":"82bf2b352e7984d0e3c23e729aa01a6f521cd2ff13f5531367d9462d5e493d54"} Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.964672 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6" event={"ID":"ec05f4fb-ca34-4f68-b94e-95b08b21d57b","Type":"ContainerStarted","Data":"ecb1220c0dabcc1b66c12a5ac29cdc2d658c4f6d7e4e3b13d7d8ffefbac7d116"} Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.967794 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-qrr4f" event={"ID":"8a4f09fc-1405-4a74-be87-1b94ee06742a","Type":"ContainerStarted","Data":"8ba61bb75ac112f3db6a23e8789bbfc4a5eb05eaa6031eaaabb5f4d756f8d3eb"} Nov 24 17:40:39 crc kubenswrapper[4808]: E1124 17:40:39.969416 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6" podUID="ec05f4fb-ca34-4f68-b94e-95b08b21d57b" Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.970622 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-66sz4" event={"ID":"eb8cf6b6-33aa-427e-9d07-9f95c58c7d30","Type":"ContainerStarted","Data":"a0397bf8bd8ac515387bf46afaf0f987622910a14e22dba0943843b01b48d393"} Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.973716 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-8659db65fc-r4hcb" event={"ID":"01054a01-64a1-4c81-b03f-5dfdf55ee4cf","Type":"ContainerStarted","Data":"527f4987f1649e345fe6314ffefee9066be0753be0b715459ed809a10e564764"} Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.991434 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mdswb" event={"ID":"2c56d9f3-3ab5-4f3c-b250-e926807dbfc3","Type":"ContainerStarted","Data":"bdfb76db3ba81fe342f65a94d101fb453cdfa3f0ea8850b4a201498b02669af8"} Nov 24 17:40:39 crc kubenswrapper[4808]: I1124 17:40:39.998342 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q5xs5" event={"ID":"58d593d8-712e-4eb5-8b1c-5df55db3f1d9","Type":"ContainerStarted","Data":"33a0e21ec39ef68e735da950f4d876ca71f6775b93fa32c06901b0ac3c95bb08"} Nov 24 17:40:40 crc kubenswrapper[4808]: I1124 17:40:40.000746 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-b5vdf" event={"ID":"af79a806-a6fb-40a1-8ac9-3fccd532c1fc","Type":"ContainerStarted","Data":"383a3b0b380560e79e6b42b7537d2593cc78394350126ed7b6e441b4b08e9c51"} Nov 24 17:40:40 crc kubenswrapper[4808]: I1124 17:40:40.017573 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-rcwzz" event={"ID":"47c249db-b332-44cf-baa5-811e04500c0f","Type":"ContainerStarted","Data":"0fa4aba8853e02a8bb4cd5e86ee4dbdb00140408f3140af446ed8890902612e2"} Nov 24 17:40:40 crc kubenswrapper[4808]: I1124 17:40:40.027623 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-668bd84cc7-lm5c4" event={"ID":"dc05be1a-a87d-4829-b2a8-c9d454b7d343","Type":"ContainerStarted","Data":"a8ea97067a50aac432072fbd803eeb15c61f510841177171a6ce79d46196309c"} Nov 24 17:40:40 crc kubenswrapper[4808]: I1124 17:40:40.033341 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-942s7" event={"ID":"bf340213-3bb6-4fc3-b8ed-9728393e9a5a","Type":"ContainerStarted","Data":"dcd4ae76df8b901004cb535d6bc58ce35122514f6b86d1ee172a9d3ce635a174"} Nov 24 17:40:40 crc kubenswrapper[4808]: I1124 17:40:40.040693 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4r97r" event={"ID":"5d6ac280-f3af-4ee8-8933-928ed0c501fc","Type":"ContainerStarted","Data":"5cbe57274d23b7f230c8cdf1345e64013b2f5c2d57f7204e7cff4dd3cc34179b"} Nov 24 17:40:40 crc kubenswrapper[4808]: E1124 17:40:40.044199 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4r97r" podUID="5d6ac280-f3af-4ee8-8933-928ed0c501fc" Nov 24 17:40:40 crc kubenswrapper[4808]: I1124 17:40:40.053234 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh" event={"ID":"c62f31fd-04d1-48cc-ae7a-98e1e64c5af7","Type":"ContainerStarted","Data":"64f548c7811af6d1dd9267611b064a385d5d3a7beadff5d2211350f1befabccd"} Nov 24 17:40:40 crc kubenswrapper[4808]: I1124 17:40:40.061632 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn" event={"ID":"321df5bf-732b-4982-841b-3a9a0ddb73f2","Type":"ContainerStarted","Data":"9df087d14e515f16d3298fdc60b144096fbd3b6eb5ed82069cf0b515900c9bd2"} Nov 24 17:40:40 crc kubenswrapper[4808]: I1124 17:40:40.069205 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tg4qk" event={"ID":"9d1d2914-59b8-4660-80a4-d2a1cc17c2bc","Type":"ContainerStarted","Data":"d93f1e14e4d84d848bef9b9c2e6143c79ff6212d1358eb245b0327748f39239c"} Nov 24 17:40:40 crc kubenswrapper[4808]: E1124 17:40:40.070259 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh" podUID="c62f31fd-04d1-48cc-ae7a-98e1e64c5af7" Nov 24 17:40:40 crc kubenswrapper[4808]: E1124 17:40:40.071050 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn" podUID="321df5bf-732b-4982-841b-3a9a0ddb73f2" Nov 24 17:40:40 crc kubenswrapper[4808]: I1124 17:40:40.072465 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" event={"ID":"58b1610a-0122-447e-a0a0-c2cb927a571f","Type":"ContainerStarted","Data":"74ea2d697e1a834d6f517a0f4011bf7069bfd1f97e2038e9fce299c4eae64a38"} Nov 24 17:40:40 crc kubenswrapper[4808]: E1124 17:40:40.115473 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" podUID="58b1610a-0122-447e-a0a0-c2cb927a571f" Nov 24 17:40:41 crc kubenswrapper[4808]: I1124 17:40:41.082253 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" event={"ID":"c796b113-a891-48b2-9a11-50a2577baf7e","Type":"ContainerStarted","Data":"d589ea2d06cab1c7a7ed1605d0e61baadbc0abc8f9da58ab59b372123a764b2d"} Nov 24 17:40:41 crc kubenswrapper[4808]: E1124 17:40:41.086171 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4r97r" podUID="5d6ac280-f3af-4ee8-8933-928ed0c501fc" Nov 24 17:40:41 crc kubenswrapper[4808]: E1124 17:40:41.087183 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" podUID="58b1610a-0122-447e-a0a0-c2cb927a571f" Nov 24 17:40:41 crc kubenswrapper[4808]: E1124 17:40:41.087913 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh" podUID="c62f31fd-04d1-48cc-ae7a-98e1e64c5af7" Nov 24 17:40:41 crc kubenswrapper[4808]: E1124 17:40:41.088167 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn" podUID="321df5bf-732b-4982-841b-3a9a0ddb73f2" Nov 24 17:40:41 crc kubenswrapper[4808]: E1124 17:40:41.088517 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6" podUID="ec05f4fb-ca34-4f68-b94e-95b08b21d57b" Nov 24 17:40:41 crc kubenswrapper[4808]: I1124 17:40:41.643776 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-metrics-certs\") pod \"openstack-operator-controller-manager-79c6ffd6b7-vvfcw\" (UID: \"e597e7fc-3b56-493b-bcda-10669936636a\") " pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:41 crc kubenswrapper[4808]: I1124 17:40:41.665974 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e597e7fc-3b56-493b-bcda-10669936636a-metrics-certs\") pod \"openstack-operator-controller-manager-79c6ffd6b7-vvfcw\" (UID: \"e597e7fc-3b56-493b-bcda-10669936636a\") " pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:41 crc kubenswrapper[4808]: I1124 17:40:41.853312 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:50 crc kubenswrapper[4808]: I1124 17:40:50.566185 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw"] Nov 24 17:40:51 crc kubenswrapper[4808]: I1124 17:40:51.151186 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" event={"ID":"e597e7fc-3b56-493b-bcda-10669936636a","Type":"ContainerStarted","Data":"b244dabc74587fbd935041a41fd3f44ed87a017b8ea0f003977dfa55e7880653"} Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.163305 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-nv6j9" event={"ID":"edd12c74-ba4f-4cb4-8531-b98607fdab41","Type":"ContainerStarted","Data":"f30fc60e7aee4e94fb393a603c6ce56e0f4583bb66c588af240882409942f2ba"} Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.169439 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tg4qk" event={"ID":"9d1d2914-59b8-4660-80a4-d2a1cc17c2bc","Type":"ContainerStarted","Data":"1242e68caad92ce65b4410f1277c1d5fa0931cab2403bb227a08d1fbfaca2f31"} Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.171301 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7fd7485997-ssvg7" event={"ID":"fa1e211d-28e9-4ea1-beb5-6e3d0f5b8b9a","Type":"ContainerStarted","Data":"ba2272f3832ad3f82fd1ff65d711deb7dbc2a2a558c837a7bbd4fffe368fd2a2"} Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.176605 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" event={"ID":"e597e7fc-3b56-493b-bcda-10669936636a","Type":"ContainerStarted","Data":"3dc3bf36a8cbf5dad320f4c234643e95f95e91a4355672192c6791fca7e5ad28"} Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.176824 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.181248 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-bbdb9b8fd-hfwr9" event={"ID":"2ece0883-b2fd-4309-b4bd-7639d95445e2","Type":"ContainerStarted","Data":"8ecc4fb54309efbe1235d29638b8f30c4cd1c66a6606451fe785744ae40ce144"} Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.184479 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-b5vdf" event={"ID":"af79a806-a6fb-40a1-8ac9-3fccd532c1fc","Type":"ContainerStarted","Data":"5ab8acd33e8fe6e6676426eaf016fb0bb505e0f4c3969ec3e6174be8ab9a2ebf"} Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.189983 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q5xs5" event={"ID":"58d593d8-712e-4eb5-8b1c-5df55db3f1d9","Type":"ContainerStarted","Data":"55394f0756c38f73530d141baca535a1afeb9a194df67f71624973f9b8fa3019"} Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.194585 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-66sz4" event={"ID":"eb8cf6b6-33aa-427e-9d07-9f95c58c7d30","Type":"ContainerStarted","Data":"4c0c0b2517987fe39cb27119aafce9666f9abc0ebd2d64a3f3f3f0b748a8f7aa"} Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.199812 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-8659db65fc-r4hcb" event={"ID":"01054a01-64a1-4c81-b03f-5dfdf55ee4cf","Type":"ContainerStarted","Data":"f0c745a6981ace70fcb51bc2c178107b903d5bb552b5f6ab305f0b84978595cb"} Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.205436 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-qrr4f" event={"ID":"8a4f09fc-1405-4a74-be87-1b94ee06742a","Type":"ContainerStarted","Data":"f601b9b0c59221ca69da7315ae90514a0de85bf0bb66a344595417cf7c177c82"} Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.207707 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5f598f884b-xqgvh" event={"ID":"ac1adbdb-b43a-4d5a-b599-7e2993f72a86","Type":"ContainerStarted","Data":"0c0a17f0bf42a23cb01897c3711b1704a8c2f37c02f89e213c0fca259e14633c"} Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.213673 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-668bd84cc7-lm5c4" event={"ID":"dc05be1a-a87d-4829-b2a8-c9d454b7d343","Type":"ContainerStarted","Data":"a121e0bc2f205b0592e64097845cf9a2e2b5f2d6fcef6e0a5fe7d4af86f1885d"} Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.215501 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-942s7" event={"ID":"bf340213-3bb6-4fc3-b8ed-9728393e9a5a","Type":"ContainerStarted","Data":"319505ed31f2e314f2a6f5878113131f988fd8308b42b50109f75f002a638962"} Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.217424 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-758c497787-nnb8j" event={"ID":"55559441-f0b0-4353-bbcb-55342cf6ef53","Type":"ContainerStarted","Data":"822b1b563572de596f7283add99ddbaf464b339a797c0021b9bac1073bef3b41"} Nov 24 17:40:52 crc kubenswrapper[4808]: I1124 17:40:52.260812 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" podStartSLOduration=15.260795769 podStartE2EDuration="15.260795769s" podCreationTimestamp="2025-11-24 17:40:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:40:52.256517445 +0000 UTC m=+844.854185267" watchObservedRunningTime="2025-11-24 17:40:52.260795769 +0000 UTC m=+844.858463571" Nov 24 17:40:54 crc kubenswrapper[4808]: I1124 17:40:54.234397 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mdswb" event={"ID":"2c56d9f3-3ab5-4f3c-b250-e926807dbfc3","Type":"ContainerStarted","Data":"b6679231bec45915751e0495ddffaab0abb60081d178e8e28ec050863b0c0c36"} Nov 24 17:40:55 crc kubenswrapper[4808]: I1124 17:40:55.243510 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-rcwzz" event={"ID":"47c249db-b332-44cf-baa5-811e04500c0f","Type":"ContainerStarted","Data":"cdf5f54188d8fd96d2873719583f5caa1396e1708073c6700a41afbc74d25075"} Nov 24 17:40:55 crc kubenswrapper[4808]: I1124 17:40:55.245163 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" event={"ID":"c796b113-a891-48b2-9a11-50a2577baf7e","Type":"ContainerStarted","Data":"edb6f90b01c8209d5f168573bada7f0e3f82a1b6f93da6e85d800576d6b5e572"} Nov 24 17:40:55 crc kubenswrapper[4808]: I1124 17:40:55.247354 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5f598f884b-xqgvh" event={"ID":"ac1adbdb-b43a-4d5a-b599-7e2993f72a86","Type":"ContainerStarted","Data":"e1089783fa7f0e478074611c7e66409bb017a80c236dc997ff56332a35916c2d"} Nov 24 17:40:56 crc kubenswrapper[4808]: E1124 17:40:56.239972 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mdswb" podUID="2c56d9f3-3ab5-4f3c-b250-e926807dbfc3" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.277284 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-668bd84cc7-lm5c4" event={"ID":"dc05be1a-a87d-4829-b2a8-c9d454b7d343","Type":"ContainerStarted","Data":"2209a40e52703461fc9f30efbcb8563f246f6bb5c736d74c4118a2bc67b3912e"} Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.277914 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-668bd84cc7-lm5c4" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.283177 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-668bd84cc7-lm5c4" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.295424 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-942s7" event={"ID":"bf340213-3bb6-4fc3-b8ed-9728393e9a5a","Type":"ContainerStarted","Data":"c463e45efcc1d5839cc5cc21d2b1c5c893fd14f92b2cef5178ab8e4cd8218872"} Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.295465 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-942s7" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.296204 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-942s7" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.299551 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-668bd84cc7-lm5c4" podStartSLOduration=8.817084456 podStartE2EDuration="20.29953219s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.287337026 +0000 UTC m=+831.885004828" lastFinishedPulling="2025-11-24 17:40:50.76978476 +0000 UTC m=+843.367452562" observedRunningTime="2025-11-24 17:40:56.296355807 +0000 UTC m=+848.894023619" watchObservedRunningTime="2025-11-24 17:40:56.29953219 +0000 UTC m=+848.897199992" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.305880 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q5xs5" event={"ID":"58d593d8-712e-4eb5-8b1c-5df55db3f1d9","Type":"ContainerStarted","Data":"b7ede0623efcae11fb27f4fe9bad3c82c58c3c3a3b506910f17f7d2fa2d6ce8a"} Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.306943 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q5xs5" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.311219 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q5xs5" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.324488 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.324526 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-942s7" podStartSLOduration=9.184255564 podStartE2EDuration="20.324497986s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:38.877313436 +0000 UTC m=+831.474981238" lastFinishedPulling="2025-11-24 17:40:50.017555858 +0000 UTC m=+842.615223660" observedRunningTime="2025-11-24 17:40:56.319491461 +0000 UTC m=+848.917159263" watchObservedRunningTime="2025-11-24 17:40:56.324497986 +0000 UTC m=+848.922165778" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.332412 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-758c497787-nnb8j" event={"ID":"55559441-f0b0-4353-bbcb-55342cf6ef53","Type":"ContainerStarted","Data":"c86fdbfbd9de29649a661ca5d3f6b750b2fb1621ff0526ede3aeab72eb657a2d"} Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.333680 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-758c497787-nnb8j" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.341885 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tg4qk" event={"ID":"9d1d2914-59b8-4660-80a4-d2a1cc17c2bc","Type":"ContainerStarted","Data":"496bf28a43238909e1e0a83763faabc74fb5070a3eab59b8b5c9766e608a66ad"} Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.342142 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tg4qk" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.349503 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-758c497787-nnb8j" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.349859 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tg4qk" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.383383 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2fzkr" event={"ID":"352d8a11-a201-4bc3-8d10-8378a5e04bfb","Type":"ContainerStarted","Data":"632931792cac9fdb9d792413c4f65c3e14d035ec0f9e6bb9e8c1729f65855e97"} Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.393966 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-758c497787-nnb8j" podStartSLOduration=9.086497344 podStartE2EDuration="20.393941137s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:38.805133274 +0000 UTC m=+831.402801076" lastFinishedPulling="2025-11-24 17:40:50.112577057 +0000 UTC m=+842.710244869" observedRunningTime="2025-11-24 17:40:56.383337508 +0000 UTC m=+848.981005310" watchObservedRunningTime="2025-11-24 17:40:56.393941137 +0000 UTC m=+848.991608939" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.394236 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-66sz4" event={"ID":"eb8cf6b6-33aa-427e-9d07-9f95c58c7d30","Type":"ContainerStarted","Data":"34e60e79ff0fd91209325e9ce3ba811564ca75fa90b5a700c05e75ef7d3211a9"} Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.395361 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-66sz4" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.419797 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-66sz4" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.435377 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-8659db65fc-r4hcb" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.439872 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-8659db65fc-r4hcb" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.449050 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-bbdb9b8fd-hfwr9" event={"ID":"2ece0883-b2fd-4309-b4bd-7639d95445e2","Type":"ContainerStarted","Data":"98b1f05f9e7f8b681d78cea3e0b1f8627919df5b9bba3cb37344dd7ecfd9c594"} Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.450679 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-bbdb9b8fd-hfwr9" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.499347 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-bbdb9b8fd-hfwr9" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.510520 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-nv6j9" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.514235 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q5xs5" podStartSLOduration=9.587978721 podStartE2EDuration="20.514203045s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.187579457 +0000 UTC m=+831.785247259" lastFinishedPulling="2025-11-24 17:40:50.113803781 +0000 UTC m=+842.711471583" observedRunningTime="2025-11-24 17:40:56.512963649 +0000 UTC m=+849.110631451" watchObservedRunningTime="2025-11-24 17:40:56.514203045 +0000 UTC m=+849.111870867" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.519845 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-nv6j9" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.527196 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-qrr4f" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.527944 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" podStartSLOduration=10.367380366 podStartE2EDuration="20.527930115s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.952117031 +0000 UTC m=+832.549784833" lastFinishedPulling="2025-11-24 17:40:50.11266672 +0000 UTC m=+842.710334582" observedRunningTime="2025-11-24 17:40:56.442815048 +0000 UTC m=+849.040482860" watchObservedRunningTime="2025-11-24 17:40:56.527930115 +0000 UTC m=+849.125597917" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.543272 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-qrr4f" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.544769 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-rcwzz" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.551207 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-b5vdf" event={"ID":"af79a806-a6fb-40a1-8ac9-3fccd532c1fc","Type":"ContainerStarted","Data":"811bfb66f13ab6243c8923e9632307b40ac3b273ebdad622dc599786b05648b2"} Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.554641 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-b5vdf" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.573161 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-b5vdf" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.582006 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tg4qk" podStartSLOduration=9.73793233 podStartE2EDuration="20.581983917s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.173745107 +0000 UTC m=+831.771412909" lastFinishedPulling="2025-11-24 17:40:50.017796684 +0000 UTC m=+842.615464496" observedRunningTime="2025-11-24 17:40:56.539302876 +0000 UTC m=+849.136970678" watchObservedRunningTime="2025-11-24 17:40:56.581983917 +0000 UTC m=+849.179651719" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.586556 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-bbdb9b8fd-hfwr9" podStartSLOduration=9.009510874 podStartE2EDuration="20.58653097s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:38.535845219 +0000 UTC m=+831.133513021" lastFinishedPulling="2025-11-24 17:40:50.112865305 +0000 UTC m=+842.710533117" observedRunningTime="2025-11-24 17:40:56.561479571 +0000 UTC m=+849.159147393" watchObservedRunningTime="2025-11-24 17:40:56.58653097 +0000 UTC m=+849.184198772" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.594848 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7fd7485997-ssvg7" event={"ID":"fa1e211d-28e9-4ea1-beb5-6e3d0f5b8b9a","Type":"ContainerStarted","Data":"7af087e14502eabfb0fb01336b997436ba0619a23024e91f2b20a4d52a589eb9"} Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.594917 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5f598f884b-xqgvh" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.596345 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7fd7485997-ssvg7" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.596869 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-66sz4" podStartSLOduration=9.796295447 podStartE2EDuration="20.59685686s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.312127017 +0000 UTC m=+831.909794819" lastFinishedPulling="2025-11-24 17:40:50.11268843 +0000 UTC m=+842.710356232" observedRunningTime="2025-11-24 17:40:56.589131525 +0000 UTC m=+849.186799337" watchObservedRunningTime="2025-11-24 17:40:56.59685686 +0000 UTC m=+849.194524662" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.600302 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7fd7485997-ssvg7" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.604963 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5f598f884b-xqgvh" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.635594 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-nv6j9" podStartSLOduration=8.730917403 podStartE2EDuration="20.635576817s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:37.95980884 +0000 UTC m=+830.557476642" lastFinishedPulling="2025-11-24 17:40:49.864468264 +0000 UTC m=+842.462136056" observedRunningTime="2025-11-24 17:40:56.634350441 +0000 UTC m=+849.232018243" watchObservedRunningTime="2025-11-24 17:40:56.635576817 +0000 UTC m=+849.233244619" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.639900 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-qrr4f" podStartSLOduration=9.043356395 podStartE2EDuration="20.639883112s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.173224272 +0000 UTC m=+831.770892074" lastFinishedPulling="2025-11-24 17:40:50.769750989 +0000 UTC m=+843.367418791" observedRunningTime="2025-11-24 17:40:56.615245635 +0000 UTC m=+849.212913437" watchObservedRunningTime="2025-11-24 17:40:56.639883112 +0000 UTC m=+849.237550914" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.652095 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-8659db65fc-r4hcb" podStartSLOduration=9.824708808 podStartE2EDuration="20.652074307s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.032164669 +0000 UTC m=+831.629832471" lastFinishedPulling="2025-11-24 17:40:49.859530168 +0000 UTC m=+842.457197970" observedRunningTime="2025-11-24 17:40:56.650441359 +0000 UTC m=+849.248109161" watchObservedRunningTime="2025-11-24 17:40:56.652074307 +0000 UTC m=+849.249742109" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.681597 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-rcwzz" podStartSLOduration=9.865556356999999 podStartE2EDuration="20.681574845s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.324484216 +0000 UTC m=+831.922152018" lastFinishedPulling="2025-11-24 17:40:50.140502704 +0000 UTC m=+842.738170506" observedRunningTime="2025-11-24 17:40:56.674306723 +0000 UTC m=+849.271974515" watchObservedRunningTime="2025-11-24 17:40:56.681574845 +0000 UTC m=+849.279242647" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.695960 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7fd7485997-ssvg7" podStartSLOduration=9.159921942 podStartE2EDuration="20.695927692s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:38.481683271 +0000 UTC m=+831.079351073" lastFinishedPulling="2025-11-24 17:40:50.017689001 +0000 UTC m=+842.615356823" observedRunningTime="2025-11-24 17:40:56.695537291 +0000 UTC m=+849.293205093" watchObservedRunningTime="2025-11-24 17:40:56.695927692 +0000 UTC m=+849.293595494" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.737474 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5f598f884b-xqgvh" podStartSLOduration=9.827591995 podStartE2EDuration="20.73744941s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.202456705 +0000 UTC m=+831.800124507" lastFinishedPulling="2025-11-24 17:40:50.11231412 +0000 UTC m=+842.709981922" observedRunningTime="2025-11-24 17:40:56.724229046 +0000 UTC m=+849.321896868" watchObservedRunningTime="2025-11-24 17:40:56.73744941 +0000 UTC m=+849.335117212" Nov 24 17:40:56 crc kubenswrapper[4808]: I1124 17:40:56.742716 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-b5vdf" podStartSLOduration=9.693963725 podStartE2EDuration="20.742694903s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.063379336 +0000 UTC m=+831.661047138" lastFinishedPulling="2025-11-24 17:40:50.112110504 +0000 UTC m=+842.709778316" observedRunningTime="2025-11-24 17:40:56.739088908 +0000 UTC m=+849.336756730" watchObservedRunningTime="2025-11-24 17:40:56.742694903 +0000 UTC m=+849.340362695" Nov 24 17:40:57 crc kubenswrapper[4808]: I1124 17:40:57.605218 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2fzkr" event={"ID":"352d8a11-a201-4bc3-8d10-8378a5e04bfb","Type":"ContainerStarted","Data":"dcb5cc20ae410f6cc05b1b13699a3eac81a4c0a49e8c4a57e53bfb87b9b17d5e"} Nov 24 17:40:57 crc kubenswrapper[4808]: I1124 17:40:57.605671 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2fzkr" Nov 24 17:40:57 crc kubenswrapper[4808]: I1124 17:40:57.614768 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" event={"ID":"c796b113-a891-48b2-9a11-50a2577baf7e","Type":"ContainerStarted","Data":"bcb35609b99d3c4231af59c39dcbcce3fedbfd41baadcb26ce43088a09c1cddf"} Nov 24 17:40:57 crc kubenswrapper[4808]: I1124 17:40:57.617941 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-8659db65fc-r4hcb" event={"ID":"01054a01-64a1-4c81-b03f-5dfdf55ee4cf","Type":"ContainerStarted","Data":"1699b10f4d0f5f35ab248ab508be7b73dc72c5fdaf8e8a7444e4af104f16bcfd"} Nov 24 17:40:57 crc kubenswrapper[4808]: I1124 17:40:57.620511 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-nv6j9" event={"ID":"edd12c74-ba4f-4cb4-8531-b98607fdab41","Type":"ContainerStarted","Data":"8346659dc6bbd654c09faf634fa9c7e9ec7a6d613a84cff15ab07e2e323f5562"} Nov 24 17:40:57 crc kubenswrapper[4808]: I1124 17:40:57.628407 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-qrr4f" event={"ID":"8a4f09fc-1405-4a74-be87-1b94ee06742a","Type":"ContainerStarted","Data":"7430d1a51984044f66385bc31dc24efccf9ac16b869a40966e4544b3bcd7e169"} Nov 24 17:40:57 crc kubenswrapper[4808]: I1124 17:40:57.632226 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-rcwzz" event={"ID":"47c249db-b332-44cf-baa5-811e04500c0f","Type":"ContainerStarted","Data":"525c5d762fa6bf86f3976d21405bc595a1d8d5974cea05c3a90d0ad53bad84d1"} Nov 24 17:40:59 crc kubenswrapper[4808]: I1124 17:40:59.128727 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7d654b7b69f77rn" Nov 24 17:40:59 crc kubenswrapper[4808]: I1124 17:40:59.163289 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2fzkr" podStartSLOduration=11.539616188 podStartE2EDuration="23.163251195s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:38.495697586 +0000 UTC m=+831.093365388" lastFinishedPulling="2025-11-24 17:40:50.119332583 +0000 UTC m=+842.717000395" observedRunningTime="2025-11-24 17:40:57.625405803 +0000 UTC m=+850.223073605" watchObservedRunningTime="2025-11-24 17:40:59.163251195 +0000 UTC m=+851.760919017" Nov 24 17:41:01 crc kubenswrapper[4808]: I1124 17:41:01.861679 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-79c6ffd6b7-vvfcw" Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.681434 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mdswb" event={"ID":"2c56d9f3-3ab5-4f3c-b250-e926807dbfc3","Type":"ContainerStarted","Data":"774726c175025de5b9087f4c3bd400d6515cbbf96b51b4364596894279aa5bca"} Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.681709 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mdswb" Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.685400 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" event={"ID":"58b1610a-0122-447e-a0a0-c2cb927a571f","Type":"ContainerStarted","Data":"8478dd43b68f3dfcb217a11eef677f1bbe8c15698805382c88f164e04a33f2e1"} Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.685432 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" event={"ID":"58b1610a-0122-447e-a0a0-c2cb927a571f","Type":"ContainerStarted","Data":"0f17527c9cb833340995f3d072c64c591f128d53a7434336f3c66b14db5a34b0"} Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.685708 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.690790 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh" event={"ID":"c62f31fd-04d1-48cc-ae7a-98e1e64c5af7","Type":"ContainerStarted","Data":"3fd918eaf88b0c50418c3135b3bc9307d58849f736b9ccc589775d51ca7629a1"} Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.690831 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh" event={"ID":"c62f31fd-04d1-48cc-ae7a-98e1e64c5af7","Type":"ContainerStarted","Data":"98e8ecc908f876383e616d393d78616a4493f9fa1879b40fb587778e5562c863"} Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.691120 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh" Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.698076 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6" event={"ID":"ec05f4fb-ca34-4f68-b94e-95b08b21d57b","Type":"ContainerStarted","Data":"443f2a8ed981538962f8486528585de7c881459d595c0ad11bed121f0cf6c6a7"} Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.698153 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6" event={"ID":"ec05f4fb-ca34-4f68-b94e-95b08b21d57b","Type":"ContainerStarted","Data":"b76cb7344fee89b9ad34a9823d0275e66796020488f5b031c23af63877703d94"} Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.699146 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6" Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.701154 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4r97r" event={"ID":"5d6ac280-f3af-4ee8-8933-928ed0c501fc","Type":"ContainerStarted","Data":"ec67b5ebb87b4922c15088ab91074e52feb4bb813d4dc1770987b2d441646d9b"} Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.703074 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn" event={"ID":"321df5bf-732b-4982-841b-3a9a0ddb73f2","Type":"ContainerStarted","Data":"32e88ae9df6b056390c946fd2e00c974f283656df7ef3e9414817b7cb496f55c"} Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.703133 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn" event={"ID":"321df5bf-732b-4982-841b-3a9a0ddb73f2","Type":"ContainerStarted","Data":"3bbb09826c0e6f08e044faad029ee9a6cfab1b88148d69cfdc5bb91ff0001504"} Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.703340 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn" Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.715102 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mdswb" podStartSLOduration=4.582297904 podStartE2EDuration="26.715072459s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.481236241 +0000 UTC m=+832.078904043" lastFinishedPulling="2025-11-24 17:41:01.614010796 +0000 UTC m=+854.211678598" observedRunningTime="2025-11-24 17:41:02.710068114 +0000 UTC m=+855.307735916" watchObservedRunningTime="2025-11-24 17:41:02.715072459 +0000 UTC m=+855.312740261" Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.729236 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh" podStartSLOduration=4.600375552 podStartE2EDuration="26.729217041s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.474611519 +0000 UTC m=+832.072279321" lastFinishedPulling="2025-11-24 17:41:01.603452998 +0000 UTC m=+854.201120810" observedRunningTime="2025-11-24 17:41:02.726874703 +0000 UTC m=+855.324542525" watchObservedRunningTime="2025-11-24 17:41:02.729217041 +0000 UTC m=+855.326884843" Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.747419 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" podStartSLOduration=4.611736382 podStartE2EDuration="26.74739631s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.480918852 +0000 UTC m=+832.078586654" lastFinishedPulling="2025-11-24 17:41:01.61657878 +0000 UTC m=+854.214246582" observedRunningTime="2025-11-24 17:41:02.746321419 +0000 UTC m=+855.343989221" watchObservedRunningTime="2025-11-24 17:41:02.74739631 +0000 UTC m=+855.345064112" Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.771793 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn" podStartSLOduration=4.48438258 podStartE2EDuration="26.771776449s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.325397351 +0000 UTC m=+831.923065153" lastFinishedPulling="2025-11-24 17:41:01.61279122 +0000 UTC m=+854.210459022" observedRunningTime="2025-11-24 17:41:02.768875025 +0000 UTC m=+855.366542827" watchObservedRunningTime="2025-11-24 17:41:02.771776449 +0000 UTC m=+855.369444251" Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.794848 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6" podStartSLOduration=4.604184494 podStartE2EDuration="26.79481726s" podCreationTimestamp="2025-11-24 17:40:36 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.412830973 +0000 UTC m=+832.010498775" lastFinishedPulling="2025-11-24 17:41:01.603463739 +0000 UTC m=+854.201131541" observedRunningTime="2025-11-24 17:41:02.789389552 +0000 UTC m=+855.387057364" watchObservedRunningTime="2025-11-24 17:41:02.79481726 +0000 UTC m=+855.392485062" Nov 24 17:41:02 crc kubenswrapper[4808]: I1124 17:41:02.811054 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4r97r" podStartSLOduration=3.648188738 podStartE2EDuration="25.811009801s" podCreationTimestamp="2025-11-24 17:40:37 +0000 UTC" firstStartedPulling="2025-11-24 17:40:39.490141916 +0000 UTC m=+832.087809718" lastFinishedPulling="2025-11-24 17:41:01.652962979 +0000 UTC m=+854.250630781" observedRunningTime="2025-11-24 17:41:02.806220551 +0000 UTC m=+855.403888353" watchObservedRunningTime="2025-11-24 17:41:02.811009801 +0000 UTC m=+855.408677603" Nov 24 17:41:06 crc kubenswrapper[4808]: I1124 17:41:06.820659 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2fzkr" Nov 24 17:41:07 crc kubenswrapper[4808]: I1124 17:41:07.488657 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-rcwzz" Nov 24 17:41:07 crc kubenswrapper[4808]: I1124 17:41:07.616385 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-zgnpn" Nov 24 17:41:07 crc kubenswrapper[4808]: I1124 17:41:07.998715 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mdswb" Nov 24 17:41:08 crc kubenswrapper[4808]: I1124 17:41:08.080636 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-9wwt6" Nov 24 17:41:08 crc kubenswrapper[4808]: I1124 17:41:08.233905 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wknfh" Nov 24 17:41:08 crc kubenswrapper[4808]: I1124 17:41:08.601376 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-299sb" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.265822 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fqxh7"] Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.267906 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fqxh7" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.270654 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-wsqzd" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.271025 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.271723 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.274807 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.285701 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fqxh7"] Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.317029 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ndjz2"] Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.318541 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.320251 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.325672 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ndjz2"] Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.386837 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w555\" (UniqueName: \"kubernetes.io/projected/220e7cfc-5d33-4e98-a8b7-1cec4ae9c556-kube-api-access-8w555\") pod \"dnsmasq-dns-675f4bcbfc-fqxh7\" (UID: \"220e7cfc-5d33-4e98-a8b7-1cec4ae9c556\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fqxh7" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.387085 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/220e7cfc-5d33-4e98-a8b7-1cec4ae9c556-config\") pod \"dnsmasq-dns-675f4bcbfc-fqxh7\" (UID: \"220e7cfc-5d33-4e98-a8b7-1cec4ae9c556\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fqxh7" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.488766 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/147efb37-6aba-4963-a24f-8f9756ac3207-config\") pod \"dnsmasq-dns-78dd6ddcc-ndjz2\" (UID: \"147efb37-6aba-4963-a24f-8f9756ac3207\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.488834 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm6b8\" (UniqueName: \"kubernetes.io/projected/147efb37-6aba-4963-a24f-8f9756ac3207-kube-api-access-sm6b8\") pod \"dnsmasq-dns-78dd6ddcc-ndjz2\" (UID: \"147efb37-6aba-4963-a24f-8f9756ac3207\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.488871 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/147efb37-6aba-4963-a24f-8f9756ac3207-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-ndjz2\" (UID: \"147efb37-6aba-4963-a24f-8f9756ac3207\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.488949 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w555\" (UniqueName: \"kubernetes.io/projected/220e7cfc-5d33-4e98-a8b7-1cec4ae9c556-kube-api-access-8w555\") pod \"dnsmasq-dns-675f4bcbfc-fqxh7\" (UID: \"220e7cfc-5d33-4e98-a8b7-1cec4ae9c556\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fqxh7" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.489010 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/220e7cfc-5d33-4e98-a8b7-1cec4ae9c556-config\") pod \"dnsmasq-dns-675f4bcbfc-fqxh7\" (UID: \"220e7cfc-5d33-4e98-a8b7-1cec4ae9c556\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fqxh7" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.489961 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/220e7cfc-5d33-4e98-a8b7-1cec4ae9c556-config\") pod \"dnsmasq-dns-675f4bcbfc-fqxh7\" (UID: \"220e7cfc-5d33-4e98-a8b7-1cec4ae9c556\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fqxh7" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.517349 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w555\" (UniqueName: \"kubernetes.io/projected/220e7cfc-5d33-4e98-a8b7-1cec4ae9c556-kube-api-access-8w555\") pod \"dnsmasq-dns-675f4bcbfc-fqxh7\" (UID: \"220e7cfc-5d33-4e98-a8b7-1cec4ae9c556\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fqxh7" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.585482 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fqxh7" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.589797 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/147efb37-6aba-4963-a24f-8f9756ac3207-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-ndjz2\" (UID: \"147efb37-6aba-4963-a24f-8f9756ac3207\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.589901 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/147efb37-6aba-4963-a24f-8f9756ac3207-config\") pod \"dnsmasq-dns-78dd6ddcc-ndjz2\" (UID: \"147efb37-6aba-4963-a24f-8f9756ac3207\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.589946 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm6b8\" (UniqueName: \"kubernetes.io/projected/147efb37-6aba-4963-a24f-8f9756ac3207-kube-api-access-sm6b8\") pod \"dnsmasq-dns-78dd6ddcc-ndjz2\" (UID: \"147efb37-6aba-4963-a24f-8f9756ac3207\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.591166 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/147efb37-6aba-4963-a24f-8f9756ac3207-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-ndjz2\" (UID: \"147efb37-6aba-4963-a24f-8f9756ac3207\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.591174 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/147efb37-6aba-4963-a24f-8f9756ac3207-config\") pod \"dnsmasq-dns-78dd6ddcc-ndjz2\" (UID: \"147efb37-6aba-4963-a24f-8f9756ac3207\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.612992 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm6b8\" (UniqueName: \"kubernetes.io/projected/147efb37-6aba-4963-a24f-8f9756ac3207-kube-api-access-sm6b8\") pod \"dnsmasq-dns-78dd6ddcc-ndjz2\" (UID: \"147efb37-6aba-4963-a24f-8f9756ac3207\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" Nov 24 17:41:26 crc kubenswrapper[4808]: I1124 17:41:26.644834 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" Nov 24 17:41:27 crc kubenswrapper[4808]: I1124 17:41:27.008498 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fqxh7"] Nov 24 17:41:27 crc kubenswrapper[4808]: I1124 17:41:27.019055 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:41:27 crc kubenswrapper[4808]: I1124 17:41:27.084766 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ndjz2"] Nov 24 17:41:27 crc kubenswrapper[4808]: I1124 17:41:27.890545 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" event={"ID":"147efb37-6aba-4963-a24f-8f9756ac3207","Type":"ContainerStarted","Data":"fb0d7abb4babf44fad7a711b7b62cb1076b6bfc268772695eb8328c558920c70"} Nov 24 17:41:27 crc kubenswrapper[4808]: I1124 17:41:27.895462 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-fqxh7" event={"ID":"220e7cfc-5d33-4e98-a8b7-1cec4ae9c556","Type":"ContainerStarted","Data":"e54d5a5216f40fd2c2ea5c3b7b82740007f9169670e9f605f5cffad25d2bd470"} Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.142538 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fqxh7"] Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.174125 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vdzbv"] Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.175912 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.208274 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vdzbv"] Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.327127 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr52n\" (UniqueName: \"kubernetes.io/projected/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-kube-api-access-cr52n\") pod \"dnsmasq-dns-666b6646f7-vdzbv\" (UID: \"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98\") " pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.327253 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-config\") pod \"dnsmasq-dns-666b6646f7-vdzbv\" (UID: \"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98\") " pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.327304 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-dns-svc\") pod \"dnsmasq-dns-666b6646f7-vdzbv\" (UID: \"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98\") " pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.428236 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr52n\" (UniqueName: \"kubernetes.io/projected/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-kube-api-access-cr52n\") pod \"dnsmasq-dns-666b6646f7-vdzbv\" (UID: \"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98\") " pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.428335 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-config\") pod \"dnsmasq-dns-666b6646f7-vdzbv\" (UID: \"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98\") " pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.428399 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-dns-svc\") pod \"dnsmasq-dns-666b6646f7-vdzbv\" (UID: \"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98\") " pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.430387 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-dns-svc\") pod \"dnsmasq-dns-666b6646f7-vdzbv\" (UID: \"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98\") " pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.432872 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-config\") pod \"dnsmasq-dns-666b6646f7-vdzbv\" (UID: \"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98\") " pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.454793 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ndjz2"] Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.459989 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr52n\" (UniqueName: \"kubernetes.io/projected/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-kube-api-access-cr52n\") pod \"dnsmasq-dns-666b6646f7-vdzbv\" (UID: \"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98\") " pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.480723 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cfr84"] Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.482164 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.501322 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cfr84"] Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.516453 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.631198 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48354d67-2079-4f48-a657-e008c3084f01-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-cfr84\" (UID: \"48354d67-2079-4f48-a657-e008c3084f01\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.631287 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48354d67-2079-4f48-a657-e008c3084f01-config\") pod \"dnsmasq-dns-57d769cc4f-cfr84\" (UID: \"48354d67-2079-4f48-a657-e008c3084f01\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.631361 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8mmr\" (UniqueName: \"kubernetes.io/projected/48354d67-2079-4f48-a657-e008c3084f01-kube-api-access-r8mmr\") pod \"dnsmasq-dns-57d769cc4f-cfr84\" (UID: \"48354d67-2079-4f48-a657-e008c3084f01\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.733116 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48354d67-2079-4f48-a657-e008c3084f01-config\") pod \"dnsmasq-dns-57d769cc4f-cfr84\" (UID: \"48354d67-2079-4f48-a657-e008c3084f01\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.733493 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8mmr\" (UniqueName: \"kubernetes.io/projected/48354d67-2079-4f48-a657-e008c3084f01-kube-api-access-r8mmr\") pod \"dnsmasq-dns-57d769cc4f-cfr84\" (UID: \"48354d67-2079-4f48-a657-e008c3084f01\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.733547 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48354d67-2079-4f48-a657-e008c3084f01-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-cfr84\" (UID: \"48354d67-2079-4f48-a657-e008c3084f01\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.734639 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48354d67-2079-4f48-a657-e008c3084f01-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-cfr84\" (UID: \"48354d67-2079-4f48-a657-e008c3084f01\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.735434 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48354d67-2079-4f48-a657-e008c3084f01-config\") pod \"dnsmasq-dns-57d769cc4f-cfr84\" (UID: \"48354d67-2079-4f48-a657-e008c3084f01\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.755391 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8mmr\" (UniqueName: \"kubernetes.io/projected/48354d67-2079-4f48-a657-e008c3084f01-kube-api-access-r8mmr\") pod \"dnsmasq-dns-57d769cc4f-cfr84\" (UID: \"48354d67-2079-4f48-a657-e008c3084f01\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.826397 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" Nov 24 17:41:29 crc kubenswrapper[4808]: I1124 17:41:29.998212 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vdzbv"] Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.322451 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.323964 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.325645 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.325888 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.326138 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.326324 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.326428 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.327798 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-rkkz4" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.332991 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.359510 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.448113 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vppw8\" (UniqueName: \"kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-kube-api-access-vppw8\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.448218 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e0b9711c-4309-40aa-8902-e2ad54e47000-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.448260 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-config-data\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.448285 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.448303 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.461897 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.461964 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.461994 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.462038 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.462071 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.462094 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e0b9711c-4309-40aa-8902-e2ad54e47000-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.563751 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e0b9711c-4309-40aa-8902-e2ad54e47000-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.563901 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-config-data\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.563946 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.563967 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.564000 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.564066 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.564086 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.564144 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.564166 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.564206 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e0b9711c-4309-40aa-8902-e2ad54e47000-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.564235 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vppw8\" (UniqueName: \"kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-kube-api-access-vppw8\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.564940 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.565311 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-config-data\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.566002 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.567144 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.567997 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.568057 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.570659 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.570690 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.576769 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e0b9711c-4309-40aa-8902-e2ad54e47000-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.578982 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e0b9711c-4309-40aa-8902-e2ad54e47000-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.581826 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vppw8\" (UniqueName: \"kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-kube-api-access-vppw8\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.590499 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.621434 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.622883 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.626238 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.626265 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.626491 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.626622 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-hvrrh" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.626724 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.626816 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.626835 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.632004 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.657779 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.766955 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.767027 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.767090 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.767122 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.767155 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldj6w\" (UniqueName: \"kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-kube-api-access-ldj6w\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.767334 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.767418 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.767507 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.767565 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c03d629c-1abf-440f-b3c8-d70e93e36eb1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.767644 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c03d629c-1abf-440f-b3c8-d70e93e36eb1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.767683 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.875306 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.875352 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.875389 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldj6w\" (UniqueName: \"kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-kube-api-access-ldj6w\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.875467 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.875503 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.875562 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.875587 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c03d629c-1abf-440f-b3c8-d70e93e36eb1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.875625 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c03d629c-1abf-440f-b3c8-d70e93e36eb1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.875647 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.875680 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.875697 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.877647 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.878550 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.878874 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.879099 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.879630 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.883433 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.884290 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.886651 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c03d629c-1abf-440f-b3c8-d70e93e36eb1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.895586 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.896513 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldj6w\" (UniqueName: \"kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-kube-api-access-ldj6w\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.902641 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c03d629c-1abf-440f-b3c8-d70e93e36eb1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.909321 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:30 crc kubenswrapper[4808]: I1124 17:41:30.966062 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:41:31 crc kubenswrapper[4808]: I1124 17:41:31.899853 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:31.908885 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:31.909042 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:31.914867 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-5zqzv" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:31.915151 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:31.915179 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:31.918488 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:31.921486 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.090696 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.090757 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-config-data-default\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.090785 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-kolla-config\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.090844 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.090950 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.091081 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.091251 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.091365 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9x5l\" (UniqueName: \"kubernetes.io/projected/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-kube-api-access-f9x5l\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.192166 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.192252 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.192299 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9x5l\" (UniqueName: \"kubernetes.io/projected/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-kube-api-access-f9x5l\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.192331 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.192350 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-config-data-default\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.192370 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-kolla-config\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.192390 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.192406 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.192484 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.192808 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.193601 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-kolla-config\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.194713 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.194855 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-config-data-default\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.198964 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.204733 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.214522 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9x5l\" (UniqueName: \"kubernetes.io/projected/ea9156df-b46b-43f4-97b8-6f1aaf74b6f6-kube-api-access-f9x5l\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.218801 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6\") " pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.244779 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 17:41:32 crc kubenswrapper[4808]: I1124 17:41:32.939094 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" event={"ID":"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98","Type":"ContainerStarted","Data":"ea5cb9fe469546cced8be35a6b3f1bed64db5fdb26a434fbaf2b55e3dcc25a2d"} Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.377898 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.384357 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.387452 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-srh8k" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.387627 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.387990 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.388804 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.389094 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.514816 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq9kv\" (UniqueName: \"kubernetes.io/projected/73a668e5-4465-455b-a109-3fcee8ea831f-kube-api-access-wq9kv\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.514913 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/73a668e5-4465-455b-a109-3fcee8ea831f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.515003 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/73a668e5-4465-455b-a109-3fcee8ea831f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.515073 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.515166 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73a668e5-4465-455b-a109-3fcee8ea831f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.515392 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/73a668e5-4465-455b-a109-3fcee8ea831f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.515493 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/73a668e5-4465-455b-a109-3fcee8ea831f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.515663 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73a668e5-4465-455b-a109-3fcee8ea831f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.617183 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73a668e5-4465-455b-a109-3fcee8ea831f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.617255 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq9kv\" (UniqueName: \"kubernetes.io/projected/73a668e5-4465-455b-a109-3fcee8ea831f-kube-api-access-wq9kv\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.617312 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/73a668e5-4465-455b-a109-3fcee8ea831f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.618094 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/73a668e5-4465-455b-a109-3fcee8ea831f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.618735 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.618760 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73a668e5-4465-455b-a109-3fcee8ea831f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.618816 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/73a668e5-4465-455b-a109-3fcee8ea831f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.618859 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/73a668e5-4465-455b-a109-3fcee8ea831f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.619185 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/73a668e5-4465-455b-a109-3fcee8ea831f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.619366 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.619607 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/73a668e5-4465-455b-a109-3fcee8ea831f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.619705 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73a668e5-4465-455b-a109-3fcee8ea831f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.620350 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/73a668e5-4465-455b-a109-3fcee8ea831f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.631080 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73a668e5-4465-455b-a109-3fcee8ea831f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.633718 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/73a668e5-4465-455b-a109-3fcee8ea831f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.637755 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq9kv\" (UniqueName: \"kubernetes.io/projected/73a668e5-4465-455b-a109-3fcee8ea831f-kube-api-access-wq9kv\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.658269 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"73a668e5-4465-455b-a109-3fcee8ea831f\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.706532 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.801122 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.802548 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.804843 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.804888 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-wtvnq" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.807475 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.822702 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.922820 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhkql\" (UniqueName: \"kubernetes.io/projected/3bd73c68-0bbb-4026-b240-637e5b7c17b5-kube-api-access-nhkql\") pod \"memcached-0\" (UID: \"3bd73c68-0bbb-4026-b240-637e5b7c17b5\") " pod="openstack/memcached-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.922866 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3bd73c68-0bbb-4026-b240-637e5b7c17b5-kolla-config\") pod \"memcached-0\" (UID: \"3bd73c68-0bbb-4026-b240-637e5b7c17b5\") " pod="openstack/memcached-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.922890 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd73c68-0bbb-4026-b240-637e5b7c17b5-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3bd73c68-0bbb-4026-b240-637e5b7c17b5\") " pod="openstack/memcached-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.922921 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bd73c68-0bbb-4026-b240-637e5b7c17b5-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3bd73c68-0bbb-4026-b240-637e5b7c17b5\") " pod="openstack/memcached-0" Nov 24 17:41:33 crc kubenswrapper[4808]: I1124 17:41:33.922993 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bd73c68-0bbb-4026-b240-637e5b7c17b5-config-data\") pod \"memcached-0\" (UID: \"3bd73c68-0bbb-4026-b240-637e5b7c17b5\") " pod="openstack/memcached-0" Nov 24 17:41:34 crc kubenswrapper[4808]: I1124 17:41:34.024888 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bd73c68-0bbb-4026-b240-637e5b7c17b5-config-data\") pod \"memcached-0\" (UID: \"3bd73c68-0bbb-4026-b240-637e5b7c17b5\") " pod="openstack/memcached-0" Nov 24 17:41:34 crc kubenswrapper[4808]: I1124 17:41:34.024977 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhkql\" (UniqueName: \"kubernetes.io/projected/3bd73c68-0bbb-4026-b240-637e5b7c17b5-kube-api-access-nhkql\") pod \"memcached-0\" (UID: \"3bd73c68-0bbb-4026-b240-637e5b7c17b5\") " pod="openstack/memcached-0" Nov 24 17:41:34 crc kubenswrapper[4808]: I1124 17:41:34.025004 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3bd73c68-0bbb-4026-b240-637e5b7c17b5-kolla-config\") pod \"memcached-0\" (UID: \"3bd73c68-0bbb-4026-b240-637e5b7c17b5\") " pod="openstack/memcached-0" Nov 24 17:41:34 crc kubenswrapper[4808]: I1124 17:41:34.025045 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd73c68-0bbb-4026-b240-637e5b7c17b5-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3bd73c68-0bbb-4026-b240-637e5b7c17b5\") " pod="openstack/memcached-0" Nov 24 17:41:34 crc kubenswrapper[4808]: I1124 17:41:34.025082 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bd73c68-0bbb-4026-b240-637e5b7c17b5-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3bd73c68-0bbb-4026-b240-637e5b7c17b5\") " pod="openstack/memcached-0" Nov 24 17:41:34 crc kubenswrapper[4808]: I1124 17:41:34.025742 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bd73c68-0bbb-4026-b240-637e5b7c17b5-config-data\") pod \"memcached-0\" (UID: \"3bd73c68-0bbb-4026-b240-637e5b7c17b5\") " pod="openstack/memcached-0" Nov 24 17:41:34 crc kubenswrapper[4808]: I1124 17:41:34.026131 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3bd73c68-0bbb-4026-b240-637e5b7c17b5-kolla-config\") pod \"memcached-0\" (UID: \"3bd73c68-0bbb-4026-b240-637e5b7c17b5\") " pod="openstack/memcached-0" Nov 24 17:41:34 crc kubenswrapper[4808]: I1124 17:41:34.037314 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bd73c68-0bbb-4026-b240-637e5b7c17b5-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3bd73c68-0bbb-4026-b240-637e5b7c17b5\") " pod="openstack/memcached-0" Nov 24 17:41:34 crc kubenswrapper[4808]: I1124 17:41:34.037351 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd73c68-0bbb-4026-b240-637e5b7c17b5-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3bd73c68-0bbb-4026-b240-637e5b7c17b5\") " pod="openstack/memcached-0" Nov 24 17:41:34 crc kubenswrapper[4808]: I1124 17:41:34.049340 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhkql\" (UniqueName: \"kubernetes.io/projected/3bd73c68-0bbb-4026-b240-637e5b7c17b5-kube-api-access-nhkql\") pod \"memcached-0\" (UID: \"3bd73c68-0bbb-4026-b240-637e5b7c17b5\") " pod="openstack/memcached-0" Nov 24 17:41:34 crc kubenswrapper[4808]: I1124 17:41:34.118289 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 17:41:35 crc kubenswrapper[4808]: I1124 17:41:35.815116 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:41:35 crc kubenswrapper[4808]: I1124 17:41:35.816141 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 17:41:35 crc kubenswrapper[4808]: I1124 17:41:35.818750 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-6jvn2" Nov 24 17:41:35 crc kubenswrapper[4808]: I1124 17:41:35.826344 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:41:35 crc kubenswrapper[4808]: I1124 17:41:35.955511 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px7c8\" (UniqueName: \"kubernetes.io/projected/f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d-kube-api-access-px7c8\") pod \"kube-state-metrics-0\" (UID: \"f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d\") " pod="openstack/kube-state-metrics-0" Nov 24 17:41:36 crc kubenswrapper[4808]: I1124 17:41:36.056971 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px7c8\" (UniqueName: \"kubernetes.io/projected/f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d-kube-api-access-px7c8\") pod \"kube-state-metrics-0\" (UID: \"f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d\") " pod="openstack/kube-state-metrics-0" Nov 24 17:41:36 crc kubenswrapper[4808]: I1124 17:41:36.074243 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px7c8\" (UniqueName: \"kubernetes.io/projected/f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d-kube-api-access-px7c8\") pod \"kube-state-metrics-0\" (UID: \"f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d\") " pod="openstack/kube-state-metrics-0" Nov 24 17:41:36 crc kubenswrapper[4808]: I1124 17:41:36.141602 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 17:41:36 crc kubenswrapper[4808]: I1124 17:41:36.656424 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cfr84"] Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.338364 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zrsgl"] Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.339761 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.342041 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-scb4x" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.342230 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.342888 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.347689 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-58sfc"] Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.355592 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zrsgl"] Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.355691 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.375352 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-58sfc"] Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.424764 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-ovn-controller-tls-certs\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.424820 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfcmf\" (UniqueName: \"kubernetes.io/projected/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-kube-api-access-lfcmf\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.424904 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-var-log-ovn\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.424959 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-combined-ca-bundle\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.424986 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-var-run-ovn\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.425102 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-var-run\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.425156 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-scripts\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.526652 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-scripts\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.526718 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-ovn-controller-tls-certs\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.526748 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfcmf\" (UniqueName: \"kubernetes.io/projected/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-kube-api-access-lfcmf\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.526789 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-var-log-ovn\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.526819 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e091dc40-a06a-482a-873c-7be28cd7fd29-etc-ovs\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.526862 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e091dc40-a06a-482a-873c-7be28cd7fd29-var-lib\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.526892 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-combined-ca-bundle\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.526988 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-var-run-ovn\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.527076 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e091dc40-a06a-482a-873c-7be28cd7fd29-scripts\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.527106 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsz9k\" (UniqueName: \"kubernetes.io/projected/e091dc40-a06a-482a-873c-7be28cd7fd29-kube-api-access-fsz9k\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.527142 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-var-run\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.527163 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e091dc40-a06a-482a-873c-7be28cd7fd29-var-run\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.527307 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e091dc40-a06a-482a-873c-7be28cd7fd29-var-log\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.527381 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-var-run-ovn\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.527449 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-var-run\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.527536 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-var-log-ovn\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.529316 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-scripts\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.533463 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-ovn-controller-tls-certs\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.542805 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-combined-ca-bundle\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.543158 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfcmf\" (UniqueName: \"kubernetes.io/projected/dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec-kube-api-access-lfcmf\") pod \"ovn-controller-zrsgl\" (UID: \"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec\") " pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.629247 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e091dc40-a06a-482a-873c-7be28cd7fd29-var-lib\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.629323 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e091dc40-a06a-482a-873c-7be28cd7fd29-scripts\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.629355 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsz9k\" (UniqueName: \"kubernetes.io/projected/e091dc40-a06a-482a-873c-7be28cd7fd29-kube-api-access-fsz9k\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.629385 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e091dc40-a06a-482a-873c-7be28cd7fd29-var-run\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.629415 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e091dc40-a06a-482a-873c-7be28cd7fd29-var-log\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.629498 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e091dc40-a06a-482a-873c-7be28cd7fd29-etc-ovs\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.629663 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e091dc40-a06a-482a-873c-7be28cd7fd29-var-lib\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.629754 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e091dc40-a06a-482a-873c-7be28cd7fd29-var-run\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.629787 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e091dc40-a06a-482a-873c-7be28cd7fd29-etc-ovs\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.629911 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e091dc40-a06a-482a-873c-7be28cd7fd29-var-log\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.631926 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e091dc40-a06a-482a-873c-7be28cd7fd29-scripts\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.647174 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsz9k\" (UniqueName: \"kubernetes.io/projected/e091dc40-a06a-482a-873c-7be28cd7fd29-kube-api-access-fsz9k\") pod \"ovn-controller-ovs-58sfc\" (UID: \"e091dc40-a06a-482a-873c-7be28cd7fd29\") " pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.660326 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.684279 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.881156 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.883382 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.888448 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.888644 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.888909 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.889386 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.889408 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-f8vrh" Nov 24 17:41:39 crc kubenswrapper[4808]: I1124 17:41:39.896428 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.034788 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjg5c\" (UniqueName: \"kubernetes.io/projected/dc135c1e-57ee-4be6-835e-b9341e7226e8-kube-api-access-pjg5c\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.034848 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dc135c1e-57ee-4be6-835e-b9341e7226e8-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.034867 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc135c1e-57ee-4be6-835e-b9341e7226e8-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.034885 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc135c1e-57ee-4be6-835e-b9341e7226e8-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.034904 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.034937 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc135c1e-57ee-4be6-835e-b9341e7226e8-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.034981 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc135c1e-57ee-4be6-835e-b9341e7226e8-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.035001 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc135c1e-57ee-4be6-835e-b9341e7226e8-config\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.136274 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc135c1e-57ee-4be6-835e-b9341e7226e8-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.136330 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc135c1e-57ee-4be6-835e-b9341e7226e8-config\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.136400 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjg5c\" (UniqueName: \"kubernetes.io/projected/dc135c1e-57ee-4be6-835e-b9341e7226e8-kube-api-access-pjg5c\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.136444 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc135c1e-57ee-4be6-835e-b9341e7226e8-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.136495 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dc135c1e-57ee-4be6-835e-b9341e7226e8-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.136525 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc135c1e-57ee-4be6-835e-b9341e7226e8-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.136545 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.136573 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc135c1e-57ee-4be6-835e-b9341e7226e8-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.137221 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dc135c1e-57ee-4be6-835e-b9341e7226e8-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.137495 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc135c1e-57ee-4be6-835e-b9341e7226e8-config\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.137519 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.137830 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc135c1e-57ee-4be6-835e-b9341e7226e8-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.140143 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc135c1e-57ee-4be6-835e-b9341e7226e8-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.140642 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc135c1e-57ee-4be6-835e-b9341e7226e8-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.157321 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc135c1e-57ee-4be6-835e-b9341e7226e8-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.159526 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjg5c\" (UniqueName: \"kubernetes.io/projected/dc135c1e-57ee-4be6-835e-b9341e7226e8-kube-api-access-pjg5c\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.161173 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dc135c1e-57ee-4be6-835e-b9341e7226e8\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:40 crc kubenswrapper[4808]: I1124 17:41:40.216129 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:41 crc kubenswrapper[4808]: I1124 17:41:41.606937 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:41:42 crc kubenswrapper[4808]: I1124 17:41:42.012396 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" event={"ID":"48354d67-2079-4f48-a657-e008c3084f01","Type":"ContainerStarted","Data":"8130703c4557d3ac2cfb4ed824154860b8b9d5d4ecc0841c3ba231fd05ee0c48"} Nov 24 17:41:42 crc kubenswrapper[4808]: E1124 17:41:42.183104 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 17:41:42 crc kubenswrapper[4808]: E1124 17:41:42.183728 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sm6b8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-ndjz2_openstack(147efb37-6aba-4963-a24f-8f9756ac3207): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:41:42 crc kubenswrapper[4808]: E1124 17:41:42.184893 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" podUID="147efb37-6aba-4963-a24f-8f9756ac3207" Nov 24 17:41:42 crc kubenswrapper[4808]: E1124 17:41:42.266226 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 17:41:42 crc kubenswrapper[4808]: E1124 17:41:42.266396 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8w555,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-fqxh7_openstack(220e7cfc-5d33-4e98-a8b7-1cec4ae9c556): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:41:42 crc kubenswrapper[4808]: E1124 17:41:42.274129 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-fqxh7" podUID="220e7cfc-5d33-4e98-a8b7-1cec4ae9c556" Nov 24 17:41:42 crc kubenswrapper[4808]: I1124 17:41:42.594359 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:41:42 crc kubenswrapper[4808]: I1124 17:41:42.768601 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:41:42 crc kubenswrapper[4808]: I1124 17:41:42.776758 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 17:41:42 crc kubenswrapper[4808]: I1124 17:41:42.788421 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 17:41:42 crc kubenswrapper[4808]: I1124 17:41:42.957897 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 17:41:42 crc kubenswrapper[4808]: W1124 17:41:42.982281 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73a668e5_4465_455b_a109_3fcee8ea831f.slice/crio-e3fcfd864f1b1ba81c23827be888287873400c8ae9d7b3b2e7a2c7cdeb21e7b8 WatchSource:0}: Error finding container e3fcfd864f1b1ba81c23827be888287873400c8ae9d7b3b2e7a2c7cdeb21e7b8: Status 404 returned error can't find the container with id e3fcfd864f1b1ba81c23827be888287873400c8ae9d7b3b2e7a2c7cdeb21e7b8 Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.027905 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3bd73c68-0bbb-4026-b240-637e5b7c17b5","Type":"ContainerStarted","Data":"6ea05d68da6f5303eec66bf0089c959355f7d0b5774cd099437516c48590f214"} Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.029751 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d","Type":"ContainerStarted","Data":"d036fd37de8907c34ba864b175d890f56cfb7841ad7d08a465d2a8c855ecffaa"} Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.031839 4808 generic.go:334] "Generic (PLEG): container finished" podID="19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98" containerID="ffb8d0aba44d6f23dd8e0c9d47ae6d19684d7582cf13079f2a0026d97d3e5927" exitCode=0 Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.031870 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" event={"ID":"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98","Type":"ContainerDied","Data":"ffb8d0aba44d6f23dd8e0c9d47ae6d19684d7582cf13079f2a0026d97d3e5927"} Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.033477 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e0b9711c-4309-40aa-8902-e2ad54e47000","Type":"ContainerStarted","Data":"7eb981fb62552051609b91fd0714ad6129273494c616ef4089769ad6c9856132"} Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.034416 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zrsgl"] Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.035492 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6","Type":"ContainerStarted","Data":"ebb23bf562514b0b86cecfecf249b47c7d5155724ca8e9de15680023526e2855"} Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.036563 4808 generic.go:334] "Generic (PLEG): container finished" podID="48354d67-2079-4f48-a657-e008c3084f01" containerID="a97a4cfcf0bb2a5736f8318736807ad48310e914c548506be1f8bbc02723c4e1" exitCode=0 Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.036609 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" event={"ID":"48354d67-2079-4f48-a657-e008c3084f01","Type":"ContainerDied","Data":"a97a4cfcf0bb2a5736f8318736807ad48310e914c548506be1f8bbc02723c4e1"} Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.042517 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"73a668e5-4465-455b-a109-3fcee8ea831f","Type":"ContainerStarted","Data":"e3fcfd864f1b1ba81c23827be888287873400c8ae9d7b3b2e7a2c7cdeb21e7b8"} Nov 24 17:41:43 crc kubenswrapper[4808]: W1124 17:41:43.043855 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc7ca1ba_7fd1_40aa_8eda_8ee9a69f10ec.slice/crio-34992f914ac9139d7212501b7cd42f801869bf4ff646d64408902915e5f19760 WatchSource:0}: Error finding container 34992f914ac9139d7212501b7cd42f801869bf4ff646d64408902915e5f19760: Status 404 returned error can't find the container with id 34992f914ac9139d7212501b7cd42f801869bf4ff646d64408902915e5f19760 Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.044248 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c03d629c-1abf-440f-b3c8-d70e93e36eb1","Type":"ContainerStarted","Data":"19e8dd5c2b161f1c8ec3a27dd487466aeff7ea5927605c36949d21f883884e2c"} Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.076049 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.082229 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.084277 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.096430 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.096491 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-vdsb8" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.096752 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.097652 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.204438 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67b93485-6ca7-4192-901b-581dbf985068-config\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.208927 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/67b93485-6ca7-4192-901b-581dbf985068-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.209616 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b93485-6ca7-4192-901b-581dbf985068-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.213148 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67b93485-6ca7-4192-901b-581dbf985068-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.214923 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b93485-6ca7-4192-901b-581dbf985068-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.221628 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkjkr\" (UniqueName: \"kubernetes.io/projected/67b93485-6ca7-4192-901b-581dbf985068-kube-api-access-bkjkr\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.221721 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.221766 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b93485-6ca7-4192-901b-581dbf985068-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.248183 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.323552 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/67b93485-6ca7-4192-901b-581dbf985068-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.323663 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b93485-6ca7-4192-901b-581dbf985068-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.323734 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67b93485-6ca7-4192-901b-581dbf985068-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.323879 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b93485-6ca7-4192-901b-581dbf985068-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.323910 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkjkr\" (UniqueName: \"kubernetes.io/projected/67b93485-6ca7-4192-901b-581dbf985068-kube-api-access-bkjkr\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.323971 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.324005 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b93485-6ca7-4192-901b-581dbf985068-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.324178 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67b93485-6ca7-4192-901b-581dbf985068-config\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.325618 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/67b93485-6ca7-4192-901b-581dbf985068-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.325868 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.326213 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67b93485-6ca7-4192-901b-581dbf985068-config\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.326872 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67b93485-6ca7-4192-901b-581dbf985068-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.341430 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b93485-6ca7-4192-901b-581dbf985068-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.343936 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b93485-6ca7-4192-901b-581dbf985068-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.354159 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkjkr\" (UniqueName: \"kubernetes.io/projected/67b93485-6ca7-4192-901b-581dbf985068-kube-api-access-bkjkr\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.365123 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b93485-6ca7-4192-901b-581dbf985068-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.376081 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"67b93485-6ca7-4192-901b-581dbf985068\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.418986 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.492005 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fqxh7" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.510558 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.632866 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/147efb37-6aba-4963-a24f-8f9756ac3207-dns-svc\") pod \"147efb37-6aba-4963-a24f-8f9756ac3207\" (UID: \"147efb37-6aba-4963-a24f-8f9756ac3207\") " Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.632992 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm6b8\" (UniqueName: \"kubernetes.io/projected/147efb37-6aba-4963-a24f-8f9756ac3207-kube-api-access-sm6b8\") pod \"147efb37-6aba-4963-a24f-8f9756ac3207\" (UID: \"147efb37-6aba-4963-a24f-8f9756ac3207\") " Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.633091 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/220e7cfc-5d33-4e98-a8b7-1cec4ae9c556-config\") pod \"220e7cfc-5d33-4e98-a8b7-1cec4ae9c556\" (UID: \"220e7cfc-5d33-4e98-a8b7-1cec4ae9c556\") " Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.633123 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/147efb37-6aba-4963-a24f-8f9756ac3207-config\") pod \"147efb37-6aba-4963-a24f-8f9756ac3207\" (UID: \"147efb37-6aba-4963-a24f-8f9756ac3207\") " Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.633150 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w555\" (UniqueName: \"kubernetes.io/projected/220e7cfc-5d33-4e98-a8b7-1cec4ae9c556-kube-api-access-8w555\") pod \"220e7cfc-5d33-4e98-a8b7-1cec4ae9c556\" (UID: \"220e7cfc-5d33-4e98-a8b7-1cec4ae9c556\") " Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.633823 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/147efb37-6aba-4963-a24f-8f9756ac3207-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "147efb37-6aba-4963-a24f-8f9756ac3207" (UID: "147efb37-6aba-4963-a24f-8f9756ac3207"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.633986 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/147efb37-6aba-4963-a24f-8f9756ac3207-config" (OuterVolumeSpecName: "config") pod "147efb37-6aba-4963-a24f-8f9756ac3207" (UID: "147efb37-6aba-4963-a24f-8f9756ac3207"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.634082 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/220e7cfc-5d33-4e98-a8b7-1cec4ae9c556-config" (OuterVolumeSpecName: "config") pod "220e7cfc-5d33-4e98-a8b7-1cec4ae9c556" (UID: "220e7cfc-5d33-4e98-a8b7-1cec4ae9c556"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.634430 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/220e7cfc-5d33-4e98-a8b7-1cec4ae9c556-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.634460 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/147efb37-6aba-4963-a24f-8f9756ac3207-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.634475 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/147efb37-6aba-4963-a24f-8f9756ac3207-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.658963 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/147efb37-6aba-4963-a24f-8f9756ac3207-kube-api-access-sm6b8" (OuterVolumeSpecName: "kube-api-access-sm6b8") pod "147efb37-6aba-4963-a24f-8f9756ac3207" (UID: "147efb37-6aba-4963-a24f-8f9756ac3207"). InnerVolumeSpecName "kube-api-access-sm6b8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.659663 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/220e7cfc-5d33-4e98-a8b7-1cec4ae9c556-kube-api-access-8w555" (OuterVolumeSpecName: "kube-api-access-8w555") pod "220e7cfc-5d33-4e98-a8b7-1cec4ae9c556" (UID: "220e7cfc-5d33-4e98-a8b7-1cec4ae9c556"). InnerVolumeSpecName "kube-api-access-8w555". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.736052 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm6b8\" (UniqueName: \"kubernetes.io/projected/147efb37-6aba-4963-a24f-8f9756ac3207-kube-api-access-sm6b8\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.736117 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w555\" (UniqueName: \"kubernetes.io/projected/220e7cfc-5d33-4e98-a8b7-1cec4ae9c556-kube-api-access-8w555\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.853828 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-58sfc"] Nov 24 17:41:43 crc kubenswrapper[4808]: I1124 17:41:43.979180 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.057617 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zrsgl" event={"ID":"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec","Type":"ContainerStarted","Data":"34992f914ac9139d7212501b7cd42f801869bf4ff646d64408902915e5f19760"} Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.060558 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" event={"ID":"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98","Type":"ContainerStarted","Data":"b276eeb4e1a064392e4d0ddb9eb3d271496213051ddb1958f5d9625481893f2e"} Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.060722 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.063727 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-fqxh7" event={"ID":"220e7cfc-5d33-4e98-a8b7-1cec4ae9c556","Type":"ContainerDied","Data":"e54d5a5216f40fd2c2ea5c3b7b82740007f9169670e9f605f5cffad25d2bd470"} Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.063808 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fqxh7" Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.069351 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"dc135c1e-57ee-4be6-835e-b9341e7226e8","Type":"ContainerStarted","Data":"7119e3dd9e93e2cb2faa8afdbede775796d1e7d4ddc1b906140162310dfb6607"} Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.071051 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" event={"ID":"147efb37-6aba-4963-a24f-8f9756ac3207","Type":"ContainerDied","Data":"fb0d7abb4babf44fad7a711b7b62cb1076b6bfc268772695eb8328c558920c70"} Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.071119 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ndjz2" Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.074529 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" event={"ID":"48354d67-2079-4f48-a657-e008c3084f01","Type":"ContainerStarted","Data":"4dba849eff7b52ef59a54ddda2bba59a52dc0227de83dd3e6f234306ddebdbd8"} Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.074687 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.083550 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" podStartSLOduration=5.541771047 podStartE2EDuration="15.083529969s" podCreationTimestamp="2025-11-24 17:41:29 +0000 UTC" firstStartedPulling="2025-11-24 17:41:32.782222666 +0000 UTC m=+885.379890468" lastFinishedPulling="2025-11-24 17:41:42.323981588 +0000 UTC m=+894.921649390" observedRunningTime="2025-11-24 17:41:44.07907414 +0000 UTC m=+896.676741942" watchObservedRunningTime="2025-11-24 17:41:44.083529969 +0000 UTC m=+896.681197771" Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.101326 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" podStartSLOduration=14.002754005 podStartE2EDuration="15.101304306s" podCreationTimestamp="2025-11-24 17:41:29 +0000 UTC" firstStartedPulling="2025-11-24 17:41:41.225753046 +0000 UTC m=+893.823420848" lastFinishedPulling="2025-11-24 17:41:42.324303347 +0000 UTC m=+894.921971149" observedRunningTime="2025-11-24 17:41:44.099259867 +0000 UTC m=+896.696927689" watchObservedRunningTime="2025-11-24 17:41:44.101304306 +0000 UTC m=+896.698972108" Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.134506 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fqxh7"] Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.141384 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fqxh7"] Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.187920 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ndjz2"] Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.202104 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ndjz2"] Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.364817 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="147efb37-6aba-4963-a24f-8f9756ac3207" path="/var/lib/kubelet/pods/147efb37-6aba-4963-a24f-8f9756ac3207/volumes" Nov 24 17:41:44 crc kubenswrapper[4808]: I1124 17:41:44.365214 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="220e7cfc-5d33-4e98-a8b7-1cec4ae9c556" path="/var/lib/kubelet/pods/220e7cfc-5d33-4e98-a8b7-1cec4ae9c556/volumes" Nov 24 17:41:45 crc kubenswrapper[4808]: I1124 17:41:45.082003 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-58sfc" event={"ID":"e091dc40-a06a-482a-873c-7be28cd7fd29","Type":"ContainerStarted","Data":"2088f89b6db1e7e852d38c014010f53c456e1c3e78fbc6577b5a453b9461ee00"} Nov 24 17:41:45 crc kubenswrapper[4808]: I1124 17:41:45.083743 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"67b93485-6ca7-4192-901b-581dbf985068","Type":"ContainerStarted","Data":"6e10fc54d3d3a836eabfee91ccbe07aa4aaf4bf85b62142a0a395c4b7cfdb373"} Nov 24 17:41:49 crc kubenswrapper[4808]: I1124 17:41:49.518237 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" Nov 24 17:41:49 crc kubenswrapper[4808]: I1124 17:41:49.827995 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" Nov 24 17:41:49 crc kubenswrapper[4808]: I1124 17:41:49.878777 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vdzbv"] Nov 24 17:41:50 crc kubenswrapper[4808]: I1124 17:41:50.135488 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" podUID="19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98" containerName="dnsmasq-dns" containerID="cri-o://b276eeb4e1a064392e4d0ddb9eb3d271496213051ddb1958f5d9625481893f2e" gracePeriod=10 Nov 24 17:41:50 crc kubenswrapper[4808]: I1124 17:41:50.835470 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" Nov 24 17:41:50 crc kubenswrapper[4808]: I1124 17:41:50.968614 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-dns-svc\") pod \"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98\" (UID: \"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98\") " Nov 24 17:41:50 crc kubenswrapper[4808]: I1124 17:41:50.969092 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cr52n\" (UniqueName: \"kubernetes.io/projected/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-kube-api-access-cr52n\") pod \"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98\" (UID: \"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98\") " Nov 24 17:41:50 crc kubenswrapper[4808]: I1124 17:41:50.969122 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-config\") pod \"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98\" (UID: \"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98\") " Nov 24 17:41:50 crc kubenswrapper[4808]: I1124 17:41:50.974875 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-kube-api-access-cr52n" (OuterVolumeSpecName: "kube-api-access-cr52n") pod "19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98" (UID: "19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98"). InnerVolumeSpecName "kube-api-access-cr52n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.017990 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98" (UID: "19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.020911 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-config" (OuterVolumeSpecName: "config") pod "19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98" (UID: "19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.071082 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.071367 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cr52n\" (UniqueName: \"kubernetes.io/projected/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-kube-api-access-cr52n\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.071600 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.157296 4808 generic.go:334] "Generic (PLEG): container finished" podID="19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98" containerID="b276eeb4e1a064392e4d0ddb9eb3d271496213051ddb1958f5d9625481893f2e" exitCode=0 Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.157382 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.157368 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" event={"ID":"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98","Type":"ContainerDied","Data":"b276eeb4e1a064392e4d0ddb9eb3d271496213051ddb1958f5d9625481893f2e"} Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.157780 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vdzbv" event={"ID":"19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98","Type":"ContainerDied","Data":"ea5cb9fe469546cced8be35a6b3f1bed64db5fdb26a434fbaf2b55e3dcc25a2d"} Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.157841 4808 scope.go:117] "RemoveContainer" containerID="b276eeb4e1a064392e4d0ddb9eb3d271496213051ddb1958f5d9625481893f2e" Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.203787 4808 scope.go:117] "RemoveContainer" containerID="ffb8d0aba44d6f23dd8e0c9d47ae6d19684d7582cf13079f2a0026d97d3e5927" Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.246989 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vdzbv"] Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.257105 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vdzbv"] Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.271208 4808 scope.go:117] "RemoveContainer" containerID="b276eeb4e1a064392e4d0ddb9eb3d271496213051ddb1958f5d9625481893f2e" Nov 24 17:41:51 crc kubenswrapper[4808]: E1124 17:41:51.271748 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b276eeb4e1a064392e4d0ddb9eb3d271496213051ddb1958f5d9625481893f2e\": container with ID starting with b276eeb4e1a064392e4d0ddb9eb3d271496213051ddb1958f5d9625481893f2e not found: ID does not exist" containerID="b276eeb4e1a064392e4d0ddb9eb3d271496213051ddb1958f5d9625481893f2e" Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.271791 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b276eeb4e1a064392e4d0ddb9eb3d271496213051ddb1958f5d9625481893f2e"} err="failed to get container status \"b276eeb4e1a064392e4d0ddb9eb3d271496213051ddb1958f5d9625481893f2e\": rpc error: code = NotFound desc = could not find container \"b276eeb4e1a064392e4d0ddb9eb3d271496213051ddb1958f5d9625481893f2e\": container with ID starting with b276eeb4e1a064392e4d0ddb9eb3d271496213051ddb1958f5d9625481893f2e not found: ID does not exist" Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.271858 4808 scope.go:117] "RemoveContainer" containerID="ffb8d0aba44d6f23dd8e0c9d47ae6d19684d7582cf13079f2a0026d97d3e5927" Nov 24 17:41:51 crc kubenswrapper[4808]: E1124 17:41:51.272264 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffb8d0aba44d6f23dd8e0c9d47ae6d19684d7582cf13079f2a0026d97d3e5927\": container with ID starting with ffb8d0aba44d6f23dd8e0c9d47ae6d19684d7582cf13079f2a0026d97d3e5927 not found: ID does not exist" containerID="ffb8d0aba44d6f23dd8e0c9d47ae6d19684d7582cf13079f2a0026d97d3e5927" Nov 24 17:41:51 crc kubenswrapper[4808]: I1124 17:41:51.272309 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffb8d0aba44d6f23dd8e0c9d47ae6d19684d7582cf13079f2a0026d97d3e5927"} err="failed to get container status \"ffb8d0aba44d6f23dd8e0c9d47ae6d19684d7582cf13079f2a0026d97d3e5927\": rpc error: code = NotFound desc = could not find container \"ffb8d0aba44d6f23dd8e0c9d47ae6d19684d7582cf13079f2a0026d97d3e5927\": container with ID starting with ffb8d0aba44d6f23dd8e0c9d47ae6d19684d7582cf13079f2a0026d97d3e5927 not found: ID does not exist" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.166767 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"67b93485-6ca7-4192-901b-581dbf985068","Type":"ContainerStarted","Data":"19a53735978fdb5193941258bb7966c40196f8d9e76b9b8da9bb4c11543a01c3"} Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.169468 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6","Type":"ContainerStarted","Data":"56fec173e66d916b7eb5a779126dff700a5113f7d05e33dc1100bc8c0d680ce4"} Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.171300 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"73a668e5-4465-455b-a109-3fcee8ea831f","Type":"ContainerStarted","Data":"1dc0fd82cfd096442ce26dfd471fb91d1bb017bd082155a7c8ee5992bc89a746"} Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.172670 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e0b9711c-4309-40aa-8902-e2ad54e47000","Type":"ContainerStarted","Data":"417e38f4fcc06e5f31f575f1e377136dd887c214621183abc1e021be79950f22"} Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.174120 4808 generic.go:334] "Generic (PLEG): container finished" podID="e091dc40-a06a-482a-873c-7be28cd7fd29" containerID="b3c86c4f5b5f2062726b3077af4e4837e0c0b9ed3778f75f67410b376b4a270b" exitCode=0 Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.174193 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-58sfc" event={"ID":"e091dc40-a06a-482a-873c-7be28cd7fd29","Type":"ContainerDied","Data":"b3c86c4f5b5f2062726b3077af4e4837e0c0b9ed3778f75f67410b376b4a270b"} Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.178828 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zrsgl" event={"ID":"dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec","Type":"ContainerStarted","Data":"bdca84add7d8ff243c3313f343230b61d947a054469d6897bc688ae21c284952"} Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.179067 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-zrsgl" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.183416 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3bd73c68-0bbb-4026-b240-637e5b7c17b5","Type":"ContainerStarted","Data":"71e1fe49af7c26f3ec6cc7485fdf30f0f6758ea4f2ee7ac3ad7a6cd46bd91769"} Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.184049 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.186439 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d","Type":"ContainerStarted","Data":"ab233518ca094448e91b1460f50fe22092ce6bd95a14d785de74edc176ea346c"} Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.187041 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.188748 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"dc135c1e-57ee-4be6-835e-b9341e7226e8","Type":"ContainerStarted","Data":"b5e24da0bba111be38d021380590bced55bb156c742209b08737cfb14b9fdc93"} Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.236004 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=9.011391059 podStartE2EDuration="17.235979881s" podCreationTimestamp="2025-11-24 17:41:35 +0000 UTC" firstStartedPulling="2025-11-24 17:41:42.785070612 +0000 UTC m=+895.382738434" lastFinishedPulling="2025-11-24 17:41:51.009659454 +0000 UTC m=+903.607327256" observedRunningTime="2025-11-24 17:41:52.231212512 +0000 UTC m=+904.828880334" watchObservedRunningTime="2025-11-24 17:41:52.235979881 +0000 UTC m=+904.833647683" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.272820 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-zrsgl" podStartSLOduration=6.018010175 podStartE2EDuration="13.272799252s" podCreationTimestamp="2025-11-24 17:41:39 +0000 UTC" firstStartedPulling="2025-11-24 17:41:43.048698432 +0000 UTC m=+895.646366244" lastFinishedPulling="2025-11-24 17:41:50.303487519 +0000 UTC m=+902.901155321" observedRunningTime="2025-11-24 17:41:52.271749362 +0000 UTC m=+904.869417164" watchObservedRunningTime="2025-11-24 17:41:52.272799252 +0000 UTC m=+904.870467054" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.297117 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=11.856672391 podStartE2EDuration="19.297099219s" podCreationTimestamp="2025-11-24 17:41:33 +0000 UTC" firstStartedPulling="2025-11-24 17:41:42.789188492 +0000 UTC m=+895.386856294" lastFinishedPulling="2025-11-24 17:41:50.22961532 +0000 UTC m=+902.827283122" observedRunningTime="2025-11-24 17:41:52.293305749 +0000 UTC m=+904.890973561" watchObservedRunningTime="2025-11-24 17:41:52.297099219 +0000 UTC m=+904.894767021" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.361065 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98" path="/var/lib/kubelet/pods/19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98/volumes" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.391147 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-kp2db"] Nov 24 17:41:52 crc kubenswrapper[4808]: E1124 17:41:52.392008 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98" containerName="dnsmasq-dns" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.392109 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98" containerName="dnsmasq-dns" Nov 24 17:41:52 crc kubenswrapper[4808]: E1124 17:41:52.392230 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98" containerName="init" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.392240 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98" containerName="init" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.392439 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="19c4dba9-57dd-4d4a-b4a9-c220c7bbfe98" containerName="dnsmasq-dns" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.393219 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.403524 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.413879 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kp2db"] Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.505476 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a3ac52cd-4f6a-4360-867b-8d0a156089b4-ovs-rundir\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.505560 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3ac52cd-4f6a-4360-867b-8d0a156089b4-combined-ca-bundle\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.505589 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3ac52cd-4f6a-4360-867b-8d0a156089b4-config\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.505619 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a3ac52cd-4f6a-4360-867b-8d0a156089b4-ovn-rundir\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.505655 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkt4w\" (UniqueName: \"kubernetes.io/projected/a3ac52cd-4f6a-4360-867b-8d0a156089b4-kube-api-access-nkt4w\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.505675 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3ac52cd-4f6a-4360-867b-8d0a156089b4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.546517 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fj7c6"] Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.557210 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.557987 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fj7c6"] Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.561468 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.606663 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a3ac52cd-4f6a-4360-867b-8d0a156089b4-ovs-rundir\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.606725 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3ac52cd-4f6a-4360-867b-8d0a156089b4-combined-ca-bundle\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.606756 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3ac52cd-4f6a-4360-867b-8d0a156089b4-config\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.606785 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a3ac52cd-4f6a-4360-867b-8d0a156089b4-ovn-rundir\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.606823 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkt4w\" (UniqueName: \"kubernetes.io/projected/a3ac52cd-4f6a-4360-867b-8d0a156089b4-kube-api-access-nkt4w\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.606840 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3ac52cd-4f6a-4360-867b-8d0a156089b4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.607066 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a3ac52cd-4f6a-4360-867b-8d0a156089b4-ovs-rundir\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.607518 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a3ac52cd-4f6a-4360-867b-8d0a156089b4-ovn-rundir\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.608727 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3ac52cd-4f6a-4360-867b-8d0a156089b4-config\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.614890 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3ac52cd-4f6a-4360-867b-8d0a156089b4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.627113 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkt4w\" (UniqueName: \"kubernetes.io/projected/a3ac52cd-4f6a-4360-867b-8d0a156089b4-kube-api-access-nkt4w\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.649654 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3ac52cd-4f6a-4360-867b-8d0a156089b4-combined-ca-bundle\") pod \"ovn-controller-metrics-kp2db\" (UID: \"a3ac52cd-4f6a-4360-867b-8d0a156089b4\") " pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.708115 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-config\") pod \"dnsmasq-dns-7fd796d7df-fj7c6\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.708180 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-979cx\" (UniqueName: \"kubernetes.io/projected/e9997c22-d554-4e42-be9c-421784fa03c0-kube-api-access-979cx\") pod \"dnsmasq-dns-7fd796d7df-fj7c6\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.708560 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-fj7c6\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.708643 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-fj7c6\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.727121 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kp2db" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.757637 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fj7c6"] Nov 24 17:41:52 crc kubenswrapper[4808]: E1124 17:41:52.758209 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-979cx ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" podUID="e9997c22-d554-4e42-be9c-421784fa03c0" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.779298 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-jw8jm"] Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.781270 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.786346 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.798159 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-jw8jm"] Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.832974 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-fj7c6\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.833052 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-fj7c6\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.833122 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-config\") pod \"dnsmasq-dns-7fd796d7df-fj7c6\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.833174 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-979cx\" (UniqueName: \"kubernetes.io/projected/e9997c22-d554-4e42-be9c-421784fa03c0-kube-api-access-979cx\") pod \"dnsmasq-dns-7fd796d7df-fj7c6\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.834260 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-fj7c6\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.834368 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-config\") pod \"dnsmasq-dns-7fd796d7df-fj7c6\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.834803 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-fj7c6\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.861819 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-979cx\" (UniqueName: \"kubernetes.io/projected/e9997c22-d554-4e42-be9c-421784fa03c0-kube-api-access-979cx\") pod \"dnsmasq-dns-7fd796d7df-fj7c6\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.934500 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxqzs\" (UniqueName: \"kubernetes.io/projected/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-kube-api-access-jxqzs\") pod \"dnsmasq-dns-86db49b7ff-jw8jm\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.934569 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-jw8jm\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.934695 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-jw8jm\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.934734 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-config\") pod \"dnsmasq-dns-86db49b7ff-jw8jm\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:52 crc kubenswrapper[4808]: I1124 17:41:52.934772 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-jw8jm\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.046629 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-jw8jm\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.047320 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-config\") pod \"dnsmasq-dns-86db49b7ff-jw8jm\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.047371 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-jw8jm\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.047428 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxqzs\" (UniqueName: \"kubernetes.io/projected/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-kube-api-access-jxqzs\") pod \"dnsmasq-dns-86db49b7ff-jw8jm\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.047475 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-jw8jm\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.047689 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-jw8jm\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.048242 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-config\") pod \"dnsmasq-dns-86db49b7ff-jw8jm\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.048539 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-jw8jm\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.048540 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-jw8jm\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.065410 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxqzs\" (UniqueName: \"kubernetes.io/projected/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-kube-api-access-jxqzs\") pod \"dnsmasq-dns-86db49b7ff-jw8jm\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.155507 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.201583 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-58sfc" event={"ID":"e091dc40-a06a-482a-873c-7be28cd7fd29","Type":"ContainerStarted","Data":"25073d0c0af75d6a85c5135facf554c65a3486b800d41e06eb3277f5e83503a4"} Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.201642 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-58sfc" event={"ID":"e091dc40-a06a-482a-873c-7be28cd7fd29","Type":"ContainerStarted","Data":"4b7d9db0d6ea63b5764467291cbf7bc9bec0b24689cbb3528bd27572dda281de"} Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.201782 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.204900 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.205246 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c03d629c-1abf-440f-b3c8-d70e93e36eb1","Type":"ContainerStarted","Data":"c746e5e9913fa0751076b716717cda0a4d7faa73d4ccd89866e3c4908aef04bf"} Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.218701 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.231388 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-58sfc" podStartSLOduration=8.545982183 podStartE2EDuration="14.231367671s" podCreationTimestamp="2025-11-24 17:41:39 +0000 UTC" firstStartedPulling="2025-11-24 17:41:44.544104298 +0000 UTC m=+897.141772100" lastFinishedPulling="2025-11-24 17:41:50.229489786 +0000 UTC m=+902.827157588" observedRunningTime="2025-11-24 17:41:53.229347272 +0000 UTC m=+905.827015074" watchObservedRunningTime="2025-11-24 17:41:53.231367671 +0000 UTC m=+905.829035473" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.278469 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kp2db"] Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.350950 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-dns-svc\") pod \"e9997c22-d554-4e42-be9c-421784fa03c0\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.350989 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-ovsdbserver-nb\") pod \"e9997c22-d554-4e42-be9c-421784fa03c0\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.351106 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-config\") pod \"e9997c22-d554-4e42-be9c-421784fa03c0\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.351252 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-979cx\" (UniqueName: \"kubernetes.io/projected/e9997c22-d554-4e42-be9c-421784fa03c0-kube-api-access-979cx\") pod \"e9997c22-d554-4e42-be9c-421784fa03c0\" (UID: \"e9997c22-d554-4e42-be9c-421784fa03c0\") " Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.351548 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e9997c22-d554-4e42-be9c-421784fa03c0" (UID: "e9997c22-d554-4e42-be9c-421784fa03c0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.351561 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e9997c22-d554-4e42-be9c-421784fa03c0" (UID: "e9997c22-d554-4e42-be9c-421784fa03c0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.351843 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.351882 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.351929 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-config" (OuterVolumeSpecName: "config") pod "e9997c22-d554-4e42-be9c-421784fa03c0" (UID: "e9997c22-d554-4e42-be9c-421784fa03c0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.355540 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9997c22-d554-4e42-be9c-421784fa03c0-kube-api-access-979cx" (OuterVolumeSpecName: "kube-api-access-979cx") pod "e9997c22-d554-4e42-be9c-421784fa03c0" (UID: "e9997c22-d554-4e42-be9c-421784fa03c0"). InnerVolumeSpecName "kube-api-access-979cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.455861 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-979cx\" (UniqueName: \"kubernetes.io/projected/e9997c22-d554-4e42-be9c-421784fa03c0-kube-api-access-979cx\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.455972 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9997c22-d554-4e42-be9c-421784fa03c0-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:53 crc kubenswrapper[4808]: I1124 17:41:53.665847 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-jw8jm"] Nov 24 17:41:54 crc kubenswrapper[4808]: I1124 17:41:54.227137 4808 generic.go:334] "Generic (PLEG): container finished" podID="3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8" containerID="bf5d8ef12838a98bbe4f640bfd36ad9f3220b4a6fdb3147e2a20a41d732c40f8" exitCode=0 Nov 24 17:41:54 crc kubenswrapper[4808]: I1124 17:41:54.227262 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" event={"ID":"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8","Type":"ContainerDied","Data":"bf5d8ef12838a98bbe4f640bfd36ad9f3220b4a6fdb3147e2a20a41d732c40f8"} Nov 24 17:41:54 crc kubenswrapper[4808]: I1124 17:41:54.227562 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" event={"ID":"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8","Type":"ContainerStarted","Data":"115559c938bc8193bf85261ab8e505c009a259b406008ab2d58219625f5f41ea"} Nov 24 17:41:54 crc kubenswrapper[4808]: I1124 17:41:54.229756 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kp2db" event={"ID":"a3ac52cd-4f6a-4360-867b-8d0a156089b4","Type":"ContainerStarted","Data":"243d0e6372477801bb95ae00257912f2a14a3548eaf7ddb39896291203f26954"} Nov 24 17:41:54 crc kubenswrapper[4808]: I1124 17:41:54.229918 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-fj7c6" Nov 24 17:41:54 crc kubenswrapper[4808]: I1124 17:41:54.230513 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:41:54 crc kubenswrapper[4808]: I1124 17:41:54.311006 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fj7c6"] Nov 24 17:41:54 crc kubenswrapper[4808]: I1124 17:41:54.316611 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fj7c6"] Nov 24 17:41:54 crc kubenswrapper[4808]: I1124 17:41:54.357142 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9997c22-d554-4e42-be9c-421784fa03c0" path="/var/lib/kubelet/pods/e9997c22-d554-4e42-be9c-421784fa03c0/volumes" Nov 24 17:41:55 crc kubenswrapper[4808]: I1124 17:41:55.237174 4808 generic.go:334] "Generic (PLEG): container finished" podID="73a668e5-4465-455b-a109-3fcee8ea831f" containerID="1dc0fd82cfd096442ce26dfd471fb91d1bb017bd082155a7c8ee5992bc89a746" exitCode=0 Nov 24 17:41:55 crc kubenswrapper[4808]: I1124 17:41:55.237287 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"73a668e5-4465-455b-a109-3fcee8ea831f","Type":"ContainerDied","Data":"1dc0fd82cfd096442ce26dfd471fb91d1bb017bd082155a7c8ee5992bc89a746"} Nov 24 17:41:56 crc kubenswrapper[4808]: I1124 17:41:56.149598 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 17:41:56 crc kubenswrapper[4808]: I1124 17:41:56.246106 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"67b93485-6ca7-4192-901b-581dbf985068","Type":"ContainerStarted","Data":"db0d8cbdc922abd88e6c2067618b0d8ba73876b972dcef7938cdee40214f83cc"} Nov 24 17:41:56 crc kubenswrapper[4808]: I1124 17:41:56.248193 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"dc135c1e-57ee-4be6-835e-b9341e7226e8","Type":"ContainerStarted","Data":"84ee1bc5b0f188a357b2fceb618b8766fb93ee3cfd20a5cbbd851c7fd7f6d2ff"} Nov 24 17:41:56 crc kubenswrapper[4808]: I1124 17:41:56.250264 4808 generic.go:334] "Generic (PLEG): container finished" podID="ea9156df-b46b-43f4-97b8-6f1aaf74b6f6" containerID="56fec173e66d916b7eb5a779126dff700a5113f7d05e33dc1100bc8c0d680ce4" exitCode=0 Nov 24 17:41:56 crc kubenswrapper[4808]: I1124 17:41:56.250346 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6","Type":"ContainerDied","Data":"56fec173e66d916b7eb5a779126dff700a5113f7d05e33dc1100bc8c0d680ce4"} Nov 24 17:41:56 crc kubenswrapper[4808]: I1124 17:41:56.253214 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"73a668e5-4465-455b-a109-3fcee8ea831f","Type":"ContainerStarted","Data":"0c7e75b6e29186480765b110a990c84fd0a2cca26bf4b2cc81b77264704940c5"} Nov 24 17:41:56 crc kubenswrapper[4808]: I1124 17:41:56.261856 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" event={"ID":"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8","Type":"ContainerStarted","Data":"4b0b2422b70954cbb976ef8a394f08d8782af6e91419908a53810b9673ffdd31"} Nov 24 17:41:56 crc kubenswrapper[4808]: I1124 17:41:56.262702 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:41:56 crc kubenswrapper[4808]: I1124 17:41:56.264739 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kp2db" event={"ID":"a3ac52cd-4f6a-4360-867b-8d0a156089b4","Type":"ContainerStarted","Data":"9b6fac60501025c488ff3b5e51903351b2ae15473bd71833df0a481211b99d97"} Nov 24 17:41:56 crc kubenswrapper[4808]: I1124 17:41:56.278575 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=2.9372989670000003 podStartE2EDuration="14.278551843s" podCreationTimestamp="2025-11-24 17:41:42 +0000 UTC" firstStartedPulling="2025-11-24 17:41:44.541563334 +0000 UTC m=+897.139231136" lastFinishedPulling="2025-11-24 17:41:55.88281621 +0000 UTC m=+908.480484012" observedRunningTime="2025-11-24 17:41:56.274919438 +0000 UTC m=+908.872587240" watchObservedRunningTime="2025-11-24 17:41:56.278551843 +0000 UTC m=+908.876219655" Nov 24 17:41:56 crc kubenswrapper[4808]: I1124 17:41:56.295526 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" podStartSLOduration=4.295501296 podStartE2EDuration="4.295501296s" podCreationTimestamp="2025-11-24 17:41:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:41:56.291183121 +0000 UTC m=+908.888850933" watchObservedRunningTime="2025-11-24 17:41:56.295501296 +0000 UTC m=+908.893169099" Nov 24 17:41:56 crc kubenswrapper[4808]: I1124 17:41:56.311646 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=5.6766552919999995 podStartE2EDuration="18.311631176s" podCreationTimestamp="2025-11-24 17:41:38 +0000 UTC" firstStartedPulling="2025-11-24 17:41:43.243568312 +0000 UTC m=+895.841236114" lastFinishedPulling="2025-11-24 17:41:55.878544196 +0000 UTC m=+908.476211998" observedRunningTime="2025-11-24 17:41:56.30798259 +0000 UTC m=+908.905650392" watchObservedRunningTime="2025-11-24 17:41:56.311631176 +0000 UTC m=+908.909298978" Nov 24 17:41:56 crc kubenswrapper[4808]: I1124 17:41:56.333301 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-kp2db" podStartSLOduration=1.825883418 podStartE2EDuration="4.333283676s" podCreationTimestamp="2025-11-24 17:41:52 +0000 UTC" firstStartedPulling="2025-11-24 17:41:53.350741414 +0000 UTC m=+905.948409216" lastFinishedPulling="2025-11-24 17:41:55.858141672 +0000 UTC m=+908.455809474" observedRunningTime="2025-11-24 17:41:56.326667843 +0000 UTC m=+908.924335645" watchObservedRunningTime="2025-11-24 17:41:56.333283676 +0000 UTC m=+908.930951478" Nov 24 17:41:56 crc kubenswrapper[4808]: I1124 17:41:56.359287 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=17.301876589 podStartE2EDuration="24.359273102s" podCreationTimestamp="2025-11-24 17:41:32 +0000 UTC" firstStartedPulling="2025-11-24 17:41:42.985197605 +0000 UTC m=+895.582865407" lastFinishedPulling="2025-11-24 17:41:50.042594118 +0000 UTC m=+902.640261920" observedRunningTime="2025-11-24 17:41:56.352820144 +0000 UTC m=+908.950487946" watchObservedRunningTime="2025-11-24 17:41:56.359273102 +0000 UTC m=+908.956940904" Nov 24 17:41:57 crc kubenswrapper[4808]: I1124 17:41:57.274598 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ea9156df-b46b-43f4-97b8-6f1aaf74b6f6","Type":"ContainerStarted","Data":"23672e1c17403296dcfe491f9ed6302687d5158b077145a870299d68c6768950"} Nov 24 17:41:57 crc kubenswrapper[4808]: I1124 17:41:57.311338 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=19.803465161 podStartE2EDuration="27.31132161s" podCreationTimestamp="2025-11-24 17:41:30 +0000 UTC" firstStartedPulling="2025-11-24 17:41:42.79565159 +0000 UTC m=+895.393319392" lastFinishedPulling="2025-11-24 17:41:50.303508039 +0000 UTC m=+902.901175841" observedRunningTime="2025-11-24 17:41:57.307323054 +0000 UTC m=+909.904990856" watchObservedRunningTime="2025-11-24 17:41:57.31132161 +0000 UTC m=+909.908989412" Nov 24 17:41:58 crc kubenswrapper[4808]: I1124 17:41:58.216896 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:58 crc kubenswrapper[4808]: I1124 17:41:58.255220 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:58 crc kubenswrapper[4808]: I1124 17:41:58.281150 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:58 crc kubenswrapper[4808]: I1124 17:41:58.316887 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 24 17:41:58 crc kubenswrapper[4808]: I1124 17:41:58.420108 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:58 crc kubenswrapper[4808]: I1124 17:41:58.421821 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:58 crc kubenswrapper[4808]: I1124 17:41:58.471335 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.119852 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.325506 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.560878 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.562843 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.565519 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.565545 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.566570 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.570480 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-5cbks" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.576529 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.674241 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-scripts\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.674307 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.674374 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-config\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.674408 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.674469 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.674500 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.674536 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjfmw\" (UniqueName: \"kubernetes.io/projected/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-kube-api-access-cjfmw\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.775836 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-config\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.775895 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.775960 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.775991 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.776037 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjfmw\" (UniqueName: \"kubernetes.io/projected/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-kube-api-access-cjfmw\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.776068 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-scripts\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.776097 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.776659 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.776776 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-config\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.777603 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-scripts\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.782916 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.783299 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.792868 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.798759 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjfmw\" (UniqueName: \"kubernetes.io/projected/a8ca4d50-8de1-4d3a-b5ed-313d024d224a-kube-api-access-cjfmw\") pod \"ovn-northd-0\" (UID: \"a8ca4d50-8de1-4d3a-b5ed-313d024d224a\") " pod="openstack/ovn-northd-0" Nov 24 17:41:59 crc kubenswrapper[4808]: I1124 17:41:59.936177 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 17:42:00 crc kubenswrapper[4808]: I1124 17:42:00.395828 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 17:42:00 crc kubenswrapper[4808]: W1124 17:42:00.399751 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8ca4d50_8de1_4d3a_b5ed_313d024d224a.slice/crio-b073edd644900f7d2a6019088c1764037a2d1ddc603db062d3f22917a0f0288f WatchSource:0}: Error finding container b073edd644900f7d2a6019088c1764037a2d1ddc603db062d3f22917a0f0288f: Status 404 returned error can't find the container with id b073edd644900f7d2a6019088c1764037a2d1ddc603db062d3f22917a0f0288f Nov 24 17:42:01 crc kubenswrapper[4808]: I1124 17:42:01.313277 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a8ca4d50-8de1-4d3a-b5ed-313d024d224a","Type":"ContainerStarted","Data":"b073edd644900f7d2a6019088c1764037a2d1ddc603db062d3f22917a0f0288f"} Nov 24 17:42:02 crc kubenswrapper[4808]: I1124 17:42:02.245718 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 24 17:42:02 crc kubenswrapper[4808]: I1124 17:42:02.245776 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 24 17:42:03 crc kubenswrapper[4808]: I1124 17:42:03.157556 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:42:03 crc kubenswrapper[4808]: I1124 17:42:03.273099 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cfr84"] Nov 24 17:42:03 crc kubenswrapper[4808]: I1124 17:42:03.273391 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" podUID="48354d67-2079-4f48-a657-e008c3084f01" containerName="dnsmasq-dns" containerID="cri-o://4dba849eff7b52ef59a54ddda2bba59a52dc0227de83dd3e6f234306ddebdbd8" gracePeriod=10 Nov 24 17:42:03 crc kubenswrapper[4808]: I1124 17:42:03.707597 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 24 17:42:03 crc kubenswrapper[4808]: I1124 17:42:03.707647 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 24 17:42:03 crc kubenswrapper[4808]: I1124 17:42:03.787030 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 24 17:42:04 crc kubenswrapper[4808]: I1124 17:42:04.418846 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 24 17:42:04 crc kubenswrapper[4808]: I1124 17:42:04.827451 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" podUID="48354d67-2079-4f48-a657-e008c3084f01" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.100:5353: connect: connection refused" Nov 24 17:42:05 crc kubenswrapper[4808]: I1124 17:42:05.230799 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 24 17:42:05 crc kubenswrapper[4808]: I1124 17:42:05.308771 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 24 17:42:05 crc kubenswrapper[4808]: I1124 17:42:05.353370 4808 generic.go:334] "Generic (PLEG): container finished" podID="48354d67-2079-4f48-a657-e008c3084f01" containerID="4dba849eff7b52ef59a54ddda2bba59a52dc0227de83dd3e6f234306ddebdbd8" exitCode=0 Nov 24 17:42:05 crc kubenswrapper[4808]: I1124 17:42:05.353406 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" event={"ID":"48354d67-2079-4f48-a657-e008c3084f01","Type":"ContainerDied","Data":"4dba849eff7b52ef59a54ddda2bba59a52dc0227de83dd3e6f234306ddebdbd8"} Nov 24 17:42:05 crc kubenswrapper[4808]: I1124 17:42:05.813436 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" Nov 24 17:42:05 crc kubenswrapper[4808]: I1124 17:42:05.890268 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48354d67-2079-4f48-a657-e008c3084f01-dns-svc\") pod \"48354d67-2079-4f48-a657-e008c3084f01\" (UID: \"48354d67-2079-4f48-a657-e008c3084f01\") " Nov 24 17:42:05 crc kubenswrapper[4808]: I1124 17:42:05.890997 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48354d67-2079-4f48-a657-e008c3084f01-config\") pod \"48354d67-2079-4f48-a657-e008c3084f01\" (UID: \"48354d67-2079-4f48-a657-e008c3084f01\") " Nov 24 17:42:05 crc kubenswrapper[4808]: I1124 17:42:05.891145 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8mmr\" (UniqueName: \"kubernetes.io/projected/48354d67-2079-4f48-a657-e008c3084f01-kube-api-access-r8mmr\") pod \"48354d67-2079-4f48-a657-e008c3084f01\" (UID: \"48354d67-2079-4f48-a657-e008c3084f01\") " Nov 24 17:42:05 crc kubenswrapper[4808]: I1124 17:42:05.912511 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48354d67-2079-4f48-a657-e008c3084f01-kube-api-access-r8mmr" (OuterVolumeSpecName: "kube-api-access-r8mmr") pod "48354d67-2079-4f48-a657-e008c3084f01" (UID: "48354d67-2079-4f48-a657-e008c3084f01"). InnerVolumeSpecName "kube-api-access-r8mmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:05 crc kubenswrapper[4808]: I1124 17:42:05.939743 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48354d67-2079-4f48-a657-e008c3084f01-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "48354d67-2079-4f48-a657-e008c3084f01" (UID: "48354d67-2079-4f48-a657-e008c3084f01"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:05 crc kubenswrapper[4808]: I1124 17:42:05.967249 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48354d67-2079-4f48-a657-e008c3084f01-config" (OuterVolumeSpecName: "config") pod "48354d67-2079-4f48-a657-e008c3084f01" (UID: "48354d67-2079-4f48-a657-e008c3084f01"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:05 crc kubenswrapper[4808]: I1124 17:42:05.993526 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48354d67-2079-4f48-a657-e008c3084f01-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:05 crc kubenswrapper[4808]: I1124 17:42:05.993565 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8mmr\" (UniqueName: \"kubernetes.io/projected/48354d67-2079-4f48-a657-e008c3084f01-kube-api-access-r8mmr\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:05 crc kubenswrapper[4808]: I1124 17:42:05.993579 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48354d67-2079-4f48-a657-e008c3084f01-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.107899 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-f2jqs"] Nov 24 17:42:06 crc kubenswrapper[4808]: E1124 17:42:06.108313 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48354d67-2079-4f48-a657-e008c3084f01" containerName="dnsmasq-dns" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.108338 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="48354d67-2079-4f48-a657-e008c3084f01" containerName="dnsmasq-dns" Nov 24 17:42:06 crc kubenswrapper[4808]: E1124 17:42:06.108365 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48354d67-2079-4f48-a657-e008c3084f01" containerName="init" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.108374 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="48354d67-2079-4f48-a657-e008c3084f01" containerName="init" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.108545 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="48354d67-2079-4f48-a657-e008c3084f01" containerName="dnsmasq-dns" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.110981 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.137857 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-f2jqs"] Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.197065 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl92t\" (UniqueName: \"kubernetes.io/projected/328bde2b-812b-4cf8-b8a4-438ccbf6d108-kube-api-access-zl92t\") pod \"dnsmasq-dns-698758b865-f2jqs\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.197234 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-config\") pod \"dnsmasq-dns-698758b865-f2jqs\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.197324 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-dns-svc\") pod \"dnsmasq-dns-698758b865-f2jqs\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.197435 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-f2jqs\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.197475 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-f2jqs\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.298697 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-f2jqs\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.298809 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-f2jqs\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.298854 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl92t\" (UniqueName: \"kubernetes.io/projected/328bde2b-812b-4cf8-b8a4-438ccbf6d108-kube-api-access-zl92t\") pod \"dnsmasq-dns-698758b865-f2jqs\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.298895 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-config\") pod \"dnsmasq-dns-698758b865-f2jqs\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.298938 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-dns-svc\") pod \"dnsmasq-dns-698758b865-f2jqs\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.299932 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-config\") pod \"dnsmasq-dns-698758b865-f2jqs\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.300037 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-dns-svc\") pod \"dnsmasq-dns-698758b865-f2jqs\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.300049 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-f2jqs\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.300064 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-f2jqs\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.315459 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl92t\" (UniqueName: \"kubernetes.io/projected/328bde2b-812b-4cf8-b8a4-438ccbf6d108-kube-api-access-zl92t\") pod \"dnsmasq-dns-698758b865-f2jqs\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.362256 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a8ca4d50-8de1-4d3a-b5ed-313d024d224a","Type":"ContainerStarted","Data":"2da84b59a75be6b288b07dcf97ea883132b0f71597d5fdcbdc9c945cc9ff4f0d"} Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.362300 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a8ca4d50-8de1-4d3a-b5ed-313d024d224a","Type":"ContainerStarted","Data":"74d434b7696c3c87c9b4e1dc1885677176f1d6f52ef1abc2f57c0d230845ee00"} Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.362358 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.364801 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" event={"ID":"48354d67-2079-4f48-a657-e008c3084f01","Type":"ContainerDied","Data":"8130703c4557d3ac2cfb4ed824154860b8b9d5d4ecc0841c3ba231fd05ee0c48"} Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.364885 4808 scope.go:117] "RemoveContainer" containerID="4dba849eff7b52ef59a54ddda2bba59a52dc0227de83dd3e6f234306ddebdbd8" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.364990 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cfr84" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.387252 4808 scope.go:117] "RemoveContainer" containerID="a97a4cfcf0bb2a5736f8318736807ad48310e914c548506be1f8bbc02723c4e1" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.389879 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.090585552 podStartE2EDuration="7.389862976s" podCreationTimestamp="2025-11-24 17:41:59 +0000 UTC" firstStartedPulling="2025-11-24 17:42:00.402234435 +0000 UTC m=+912.999902237" lastFinishedPulling="2025-11-24 17:42:05.701511859 +0000 UTC m=+918.299179661" observedRunningTime="2025-11-24 17:42:06.386256011 +0000 UTC m=+918.983923813" watchObservedRunningTime="2025-11-24 17:42:06.389862976 +0000 UTC m=+918.987530778" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.403308 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cfr84"] Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.408845 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cfr84"] Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.428051 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:06 crc kubenswrapper[4808]: I1124 17:42:06.844214 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-f2jqs"] Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.222989 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.229863 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.232378 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.232623 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.233460 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-t68lg" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.233530 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.255254 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.320665 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk4hn\" (UniqueName: \"kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-kube-api-access-mk4hn\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.320731 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.320784 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-lock\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.320822 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-cache\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.320842 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.374220 4808 generic.go:334] "Generic (PLEG): container finished" podID="328bde2b-812b-4cf8-b8a4-438ccbf6d108" containerID="3ab0c0a9c8b54d6e2baff9ec27547e6367217d6c746abdf098fc7a77e022c931" exitCode=0 Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.374313 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-f2jqs" event={"ID":"328bde2b-812b-4cf8-b8a4-438ccbf6d108","Type":"ContainerDied","Data":"3ab0c0a9c8b54d6e2baff9ec27547e6367217d6c746abdf098fc7a77e022c931"} Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.374380 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-f2jqs" event={"ID":"328bde2b-812b-4cf8-b8a4-438ccbf6d108","Type":"ContainerStarted","Data":"79bff3c094f43a4e9f8f738eed6f321e084ffe6d2fbe9f2858928c51ff47d722"} Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.422072 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-lock\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.422133 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-cache\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.422166 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.422268 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk4hn\" (UniqueName: \"kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-kube-api-access-mk4hn\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.422307 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: E1124 17:42:07.422381 4808 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 17:42:07 crc kubenswrapper[4808]: E1124 17:42:07.422426 4808 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 17:42:07 crc kubenswrapper[4808]: E1124 17:42:07.422501 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift podName:c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14 nodeName:}" failed. No retries permitted until 2025-11-24 17:42:07.922459518 +0000 UTC m=+920.520127320 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift") pod "swift-storage-0" (UID: "c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14") : configmap "swift-ring-files" not found Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.422650 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.423275 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-cache\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.424557 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-lock\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.441308 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk4hn\" (UniqueName: \"kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-kube-api-access-mk4hn\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.446878 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.884042 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-2qncr"] Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.885771 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.888342 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.888608 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.901168 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.931949 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:07 crc kubenswrapper[4808]: E1124 17:42:07.932194 4808 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 17:42:07 crc kubenswrapper[4808]: E1124 17:42:07.932211 4808 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 17:42:07 crc kubenswrapper[4808]: E1124 17:42:07.932257 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift podName:c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14 nodeName:}" failed. No retries permitted until 2025-11-24 17:42:08.932242399 +0000 UTC m=+921.529910201 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift") pod "swift-storage-0" (UID: "c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14") : configmap "swift-ring-files" not found Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.944087 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-6ttz5"] Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.945718 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.949951 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-2qncr"] Nov 24 17:42:07 crc kubenswrapper[4808]: E1124 17:42:07.950669 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-cc8ps ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-cc8ps ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-2qncr" podUID="f226736e-0590-4624-a06b-fb3364774c2b" Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.955289 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-6ttz5"] Nov 24 17:42:07 crc kubenswrapper[4808]: I1124 17:42:07.975937 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-2qncr"] Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.033422 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f226736e-0590-4624-a06b-fb3364774c2b-ring-data-devices\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.033466 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f226736e-0590-4624-a06b-fb3364774c2b-scripts\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.033506 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee5479fc-5360-4541-990d-c8d1a97bae29-scripts\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.033534 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-swiftconf\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.033556 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-combined-ca-bundle\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.033580 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq55s\" (UniqueName: \"kubernetes.io/projected/ee5479fc-5360-4541-990d-c8d1a97bae29-kube-api-access-jq55s\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.033607 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc8ps\" (UniqueName: \"kubernetes.io/projected/f226736e-0590-4624-a06b-fb3364774c2b-kube-api-access-cc8ps\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.033670 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f226736e-0590-4624-a06b-fb3364774c2b-etc-swift\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.033763 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-combined-ca-bundle\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.033874 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ee5479fc-5360-4541-990d-c8d1a97bae29-ring-data-devices\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.033928 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-dispersionconf\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.033961 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ee5479fc-5360-4541-990d-c8d1a97bae29-etc-swift\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.034008 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-dispersionconf\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.034132 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-swiftconf\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.135721 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-swiftconf\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.135805 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f226736e-0590-4624-a06b-fb3364774c2b-ring-data-devices\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.135835 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f226736e-0590-4624-a06b-fb3364774c2b-scripts\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.135867 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee5479fc-5360-4541-990d-c8d1a97bae29-scripts\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.135892 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-swiftconf\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.135915 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-combined-ca-bundle\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.135939 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq55s\" (UniqueName: \"kubernetes.io/projected/ee5479fc-5360-4541-990d-c8d1a97bae29-kube-api-access-jq55s\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.135969 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc8ps\" (UniqueName: \"kubernetes.io/projected/f226736e-0590-4624-a06b-fb3364774c2b-kube-api-access-cc8ps\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.135998 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f226736e-0590-4624-a06b-fb3364774c2b-etc-swift\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.136043 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-combined-ca-bundle\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.136120 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ee5479fc-5360-4541-990d-c8d1a97bae29-ring-data-devices\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.136143 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-dispersionconf\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.136166 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ee5479fc-5360-4541-990d-c8d1a97bae29-etc-swift\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.136200 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-dispersionconf\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.136588 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f226736e-0590-4624-a06b-fb3364774c2b-ring-data-devices\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.137512 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ee5479fc-5360-4541-990d-c8d1a97bae29-etc-swift\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.137562 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f226736e-0590-4624-a06b-fb3364774c2b-scripts\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.137577 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ee5479fc-5360-4541-990d-c8d1a97bae29-ring-data-devices\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.138210 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee5479fc-5360-4541-990d-c8d1a97bae29-scripts\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.138530 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f226736e-0590-4624-a06b-fb3364774c2b-etc-swift\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.141474 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-swiftconf\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.141499 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-swiftconf\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.142258 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-combined-ca-bundle\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.143460 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-dispersionconf\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.145569 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-combined-ca-bundle\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.150426 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-dispersionconf\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.157025 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq55s\" (UniqueName: \"kubernetes.io/projected/ee5479fc-5360-4541-990d-c8d1a97bae29-kube-api-access-jq55s\") pod \"swift-ring-rebalance-6ttz5\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.162461 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc8ps\" (UniqueName: \"kubernetes.io/projected/f226736e-0590-4624-a06b-fb3364774c2b-kube-api-access-cc8ps\") pod \"swift-ring-rebalance-2qncr\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.282526 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.366634 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48354d67-2079-4f48-a657-e008c3084f01" path="/var/lib/kubelet/pods/48354d67-2079-4f48-a657-e008c3084f01/volumes" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.389418 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.389845 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-f2jqs" event={"ID":"328bde2b-812b-4cf8-b8a4-438ccbf6d108","Type":"ContainerStarted","Data":"4b54deb9d465b1902512df3a773a8a3e978f9e9fbf6ab948b077fb78427ec67a"} Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.390061 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.436670 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-f2jqs" podStartSLOduration=2.436655235 podStartE2EDuration="2.436655235s" podCreationTimestamp="2025-11-24 17:42:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:42:08.436231042 +0000 UTC m=+921.033898854" watchObservedRunningTime="2025-11-24 17:42:08.436655235 +0000 UTC m=+921.034323037" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.456871 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.542549 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f226736e-0590-4624-a06b-fb3364774c2b-ring-data-devices\") pod \"f226736e-0590-4624-a06b-fb3364774c2b\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.542613 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f226736e-0590-4624-a06b-fb3364774c2b-etc-swift\") pod \"f226736e-0590-4624-a06b-fb3364774c2b\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.542752 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cc8ps\" (UniqueName: \"kubernetes.io/projected/f226736e-0590-4624-a06b-fb3364774c2b-kube-api-access-cc8ps\") pod \"f226736e-0590-4624-a06b-fb3364774c2b\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.542834 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-swiftconf\") pod \"f226736e-0590-4624-a06b-fb3364774c2b\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.542859 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-combined-ca-bundle\") pod \"f226736e-0590-4624-a06b-fb3364774c2b\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.542895 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f226736e-0590-4624-a06b-fb3364774c2b-scripts\") pod \"f226736e-0590-4624-a06b-fb3364774c2b\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.542943 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-dispersionconf\") pod \"f226736e-0590-4624-a06b-fb3364774c2b\" (UID: \"f226736e-0590-4624-a06b-fb3364774c2b\") " Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.543185 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f226736e-0590-4624-a06b-fb3364774c2b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f226736e-0590-4624-a06b-fb3364774c2b" (UID: "f226736e-0590-4624-a06b-fb3364774c2b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.543493 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f226736e-0590-4624-a06b-fb3364774c2b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f226736e-0590-4624-a06b-fb3364774c2b" (UID: "f226736e-0590-4624-a06b-fb3364774c2b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.544089 4808 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f226736e-0590-4624-a06b-fb3364774c2b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.544107 4808 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f226736e-0590-4624-a06b-fb3364774c2b-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.545147 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f226736e-0590-4624-a06b-fb3364774c2b-scripts" (OuterVolumeSpecName: "scripts") pod "f226736e-0590-4624-a06b-fb3364774c2b" (UID: "f226736e-0590-4624-a06b-fb3364774c2b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.548678 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f226736e-0590-4624-a06b-fb3364774c2b" (UID: "f226736e-0590-4624-a06b-fb3364774c2b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.549089 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f226736e-0590-4624-a06b-fb3364774c2b" (UID: "f226736e-0590-4624-a06b-fb3364774c2b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.549143 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f226736e-0590-4624-a06b-fb3364774c2b-kube-api-access-cc8ps" (OuterVolumeSpecName: "kube-api-access-cc8ps") pod "f226736e-0590-4624-a06b-fb3364774c2b" (UID: "f226736e-0590-4624-a06b-fb3364774c2b"). InnerVolumeSpecName "kube-api-access-cc8ps". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.555314 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f226736e-0590-4624-a06b-fb3364774c2b" (UID: "f226736e-0590-4624-a06b-fb3364774c2b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.645073 4808 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.645117 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.645132 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f226736e-0590-4624-a06b-fb3364774c2b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.645143 4808 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f226736e-0590-4624-a06b-fb3364774c2b-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.645154 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cc8ps\" (UniqueName: \"kubernetes.io/projected/f226736e-0590-4624-a06b-fb3364774c2b-kube-api-access-cc8ps\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.787030 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-6ttz5"] Nov 24 17:42:08 crc kubenswrapper[4808]: I1124 17:42:08.950571 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:08 crc kubenswrapper[4808]: E1124 17:42:08.950802 4808 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 17:42:08 crc kubenswrapper[4808]: E1124 17:42:08.950817 4808 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 17:42:08 crc kubenswrapper[4808]: E1124 17:42:08.950864 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift podName:c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14 nodeName:}" failed. No retries permitted until 2025-11-24 17:42:10.950851204 +0000 UTC m=+923.548519006 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift") pod "swift-storage-0" (UID: "c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14") : configmap "swift-ring-files" not found Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.325618 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-b750-account-create-update-v7jcb"] Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.327011 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b750-account-create-update-v7jcb" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.332587 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.340103 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b750-account-create-update-v7jcb"] Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.353870 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-8z24k"] Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.355097 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8z24k" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.379418 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8z24k"] Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.405065 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6ttz5" event={"ID":"ee5479fc-5360-4541-990d-c8d1a97bae29","Type":"ContainerStarted","Data":"737eb8a706d08a39baddaded640e6d83dd3f6748855784baf344a5624ca18d00"} Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.405143 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2qncr" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.446053 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-2qncr"] Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.451444 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-2qncr"] Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.459790 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c-operator-scripts\") pod \"glance-db-create-8z24k\" (UID: \"86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c\") " pod="openstack/glance-db-create-8z24k" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.459854 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd9gc\" (UniqueName: \"kubernetes.io/projected/3563a459-e344-43c7-83ee-3da790cfe660-kube-api-access-wd9gc\") pod \"glance-b750-account-create-update-v7jcb\" (UID: \"3563a459-e344-43c7-83ee-3da790cfe660\") " pod="openstack/glance-b750-account-create-update-v7jcb" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.459911 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3563a459-e344-43c7-83ee-3da790cfe660-operator-scripts\") pod \"glance-b750-account-create-update-v7jcb\" (UID: \"3563a459-e344-43c7-83ee-3da790cfe660\") " pod="openstack/glance-b750-account-create-update-v7jcb" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.459936 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr26w\" (UniqueName: \"kubernetes.io/projected/86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c-kube-api-access-gr26w\") pod \"glance-db-create-8z24k\" (UID: \"86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c\") " pod="openstack/glance-db-create-8z24k" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.561443 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c-operator-scripts\") pod \"glance-db-create-8z24k\" (UID: \"86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c\") " pod="openstack/glance-db-create-8z24k" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.561940 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd9gc\" (UniqueName: \"kubernetes.io/projected/3563a459-e344-43c7-83ee-3da790cfe660-kube-api-access-wd9gc\") pod \"glance-b750-account-create-update-v7jcb\" (UID: \"3563a459-e344-43c7-83ee-3da790cfe660\") " pod="openstack/glance-b750-account-create-update-v7jcb" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.561980 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3563a459-e344-43c7-83ee-3da790cfe660-operator-scripts\") pod \"glance-b750-account-create-update-v7jcb\" (UID: \"3563a459-e344-43c7-83ee-3da790cfe660\") " pod="openstack/glance-b750-account-create-update-v7jcb" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.562002 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr26w\" (UniqueName: \"kubernetes.io/projected/86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c-kube-api-access-gr26w\") pod \"glance-db-create-8z24k\" (UID: \"86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c\") " pod="openstack/glance-db-create-8z24k" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.562371 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c-operator-scripts\") pod \"glance-db-create-8z24k\" (UID: \"86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c\") " pod="openstack/glance-db-create-8z24k" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.562701 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3563a459-e344-43c7-83ee-3da790cfe660-operator-scripts\") pod \"glance-b750-account-create-update-v7jcb\" (UID: \"3563a459-e344-43c7-83ee-3da790cfe660\") " pod="openstack/glance-b750-account-create-update-v7jcb" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.585096 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr26w\" (UniqueName: \"kubernetes.io/projected/86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c-kube-api-access-gr26w\") pod \"glance-db-create-8z24k\" (UID: \"86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c\") " pod="openstack/glance-db-create-8z24k" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.587916 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd9gc\" (UniqueName: \"kubernetes.io/projected/3563a459-e344-43c7-83ee-3da790cfe660-kube-api-access-wd9gc\") pod \"glance-b750-account-create-update-v7jcb\" (UID: \"3563a459-e344-43c7-83ee-3da790cfe660\") " pod="openstack/glance-b750-account-create-update-v7jcb" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.653638 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b750-account-create-update-v7jcb" Nov 24 17:42:09 crc kubenswrapper[4808]: I1124 17:42:09.672802 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8z24k" Nov 24 17:42:10 crc kubenswrapper[4808]: I1124 17:42:10.136964 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b750-account-create-update-v7jcb"] Nov 24 17:42:10 crc kubenswrapper[4808]: I1124 17:42:10.206932 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8z24k"] Nov 24 17:42:10 crc kubenswrapper[4808]: I1124 17:42:10.369190 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f226736e-0590-4624-a06b-fb3364774c2b" path="/var/lib/kubelet/pods/f226736e-0590-4624-a06b-fb3364774c2b/volumes" Nov 24 17:42:10 crc kubenswrapper[4808]: I1124 17:42:10.984667 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:10 crc kubenswrapper[4808]: E1124 17:42:10.984927 4808 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 17:42:10 crc kubenswrapper[4808]: E1124 17:42:10.984957 4808 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 17:42:10 crc kubenswrapper[4808]: E1124 17:42:10.985035 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift podName:c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14 nodeName:}" failed. No retries permitted until 2025-11-24 17:42:14.985003304 +0000 UTC m=+927.582671106 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift") pod "swift-storage-0" (UID: "c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14") : configmap "swift-ring-files" not found Nov 24 17:42:11 crc kubenswrapper[4808]: W1124 17:42:11.653498 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3563a459_e344_43c7_83ee_3da790cfe660.slice/crio-58d6363d4225f7e1beb5ead6ddfce611c3e5b3d05cd1b5887f1827dd5d4dfdd6 WatchSource:0}: Error finding container 58d6363d4225f7e1beb5ead6ddfce611c3e5b3d05cd1b5887f1827dd5d4dfdd6: Status 404 returned error can't find the container with id 58d6363d4225f7e1beb5ead6ddfce611c3e5b3d05cd1b5887f1827dd5d4dfdd6 Nov 24 17:42:12 crc kubenswrapper[4808]: I1124 17:42:12.425826 4808 generic.go:334] "Generic (PLEG): container finished" podID="3563a459-e344-43c7-83ee-3da790cfe660" containerID="d975a76363a71b34015a1762ce112417a4bde8399ede3829500729785f099064" exitCode=0 Nov 24 17:42:12 crc kubenswrapper[4808]: I1124 17:42:12.425879 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b750-account-create-update-v7jcb" event={"ID":"3563a459-e344-43c7-83ee-3da790cfe660","Type":"ContainerDied","Data":"d975a76363a71b34015a1762ce112417a4bde8399ede3829500729785f099064"} Nov 24 17:42:12 crc kubenswrapper[4808]: I1124 17:42:12.426272 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b750-account-create-update-v7jcb" event={"ID":"3563a459-e344-43c7-83ee-3da790cfe660","Type":"ContainerStarted","Data":"58d6363d4225f7e1beb5ead6ddfce611c3e5b3d05cd1b5887f1827dd5d4dfdd6"} Nov 24 17:42:12 crc kubenswrapper[4808]: I1124 17:42:12.428133 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6ttz5" event={"ID":"ee5479fc-5360-4541-990d-c8d1a97bae29","Type":"ContainerStarted","Data":"55772800d663664864db0a964b8e4d448e570041634e7cd558dc2ede4909964b"} Nov 24 17:42:12 crc kubenswrapper[4808]: I1124 17:42:12.434432 4808 generic.go:334] "Generic (PLEG): container finished" podID="86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c" containerID="0479a82d98c3c1e45a143d1dff8827a025530872bb39af91ac7b4db3e76d6ded" exitCode=0 Nov 24 17:42:12 crc kubenswrapper[4808]: I1124 17:42:12.434479 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8z24k" event={"ID":"86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c","Type":"ContainerDied","Data":"0479a82d98c3c1e45a143d1dff8827a025530872bb39af91ac7b4db3e76d6ded"} Nov 24 17:42:12 crc kubenswrapper[4808]: I1124 17:42:12.434515 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8z24k" event={"ID":"86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c","Type":"ContainerStarted","Data":"786a5ad7ccd68a05d270d505b7682f0d984375f225e42be77c919f0efea58f62"} Nov 24 17:42:12 crc kubenswrapper[4808]: I1124 17:42:12.477211 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-6ttz5" podStartSLOduration=2.519162938 podStartE2EDuration="5.477188547s" podCreationTimestamp="2025-11-24 17:42:07 +0000 UTC" firstStartedPulling="2025-11-24 17:42:08.791629862 +0000 UTC m=+921.389297664" lastFinishedPulling="2025-11-24 17:42:11.749655471 +0000 UTC m=+924.347323273" observedRunningTime="2025-11-24 17:42:12.476740084 +0000 UTC m=+925.074407886" watchObservedRunningTime="2025-11-24 17:42:12.477188547 +0000 UTC m=+925.074856339" Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.731410 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-9jp7n"] Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.740662 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9jp7n" Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.744979 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9jp7n"] Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.831411 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5213-account-create-update-q7rbp"] Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.832819 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5213-account-create-update-q7rbp" Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.832989 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sn4l\" (UniqueName: \"kubernetes.io/projected/ef8985b9-2261-40cf-bfa9-ef86fb447616-kube-api-access-2sn4l\") pod \"keystone-db-create-9jp7n\" (UID: \"ef8985b9-2261-40cf-bfa9-ef86fb447616\") " pod="openstack/keystone-db-create-9jp7n" Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.833147 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef8985b9-2261-40cf-bfa9-ef86fb447616-operator-scripts\") pod \"keystone-db-create-9jp7n\" (UID: \"ef8985b9-2261-40cf-bfa9-ef86fb447616\") " pod="openstack/keystone-db-create-9jp7n" Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.836644 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.837623 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5213-account-create-update-q7rbp"] Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.885251 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8z24k" Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.891894 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b750-account-create-update-v7jcb" Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.934843 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sn4l\" (UniqueName: \"kubernetes.io/projected/ef8985b9-2261-40cf-bfa9-ef86fb447616-kube-api-access-2sn4l\") pod \"keystone-db-create-9jp7n\" (UID: \"ef8985b9-2261-40cf-bfa9-ef86fb447616\") " pod="openstack/keystone-db-create-9jp7n" Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.934979 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fclf\" (UniqueName: \"kubernetes.io/projected/6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0-kube-api-access-2fclf\") pod \"keystone-5213-account-create-update-q7rbp\" (UID: \"6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0\") " pod="openstack/keystone-5213-account-create-update-q7rbp" Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.935036 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0-operator-scripts\") pod \"keystone-5213-account-create-update-q7rbp\" (UID: \"6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0\") " pod="openstack/keystone-5213-account-create-update-q7rbp" Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.935069 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef8985b9-2261-40cf-bfa9-ef86fb447616-operator-scripts\") pod \"keystone-db-create-9jp7n\" (UID: \"ef8985b9-2261-40cf-bfa9-ef86fb447616\") " pod="openstack/keystone-db-create-9jp7n" Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.936296 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef8985b9-2261-40cf-bfa9-ef86fb447616-operator-scripts\") pod \"keystone-db-create-9jp7n\" (UID: \"ef8985b9-2261-40cf-bfa9-ef86fb447616\") " pod="openstack/keystone-db-create-9jp7n" Nov 24 17:42:13 crc kubenswrapper[4808]: I1124 17:42:13.956398 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sn4l\" (UniqueName: \"kubernetes.io/projected/ef8985b9-2261-40cf-bfa9-ef86fb447616-kube-api-access-2sn4l\") pod \"keystone-db-create-9jp7n\" (UID: \"ef8985b9-2261-40cf-bfa9-ef86fb447616\") " pod="openstack/keystone-db-create-9jp7n" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.033712 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-fmsjw"] Nov 24 17:42:14 crc kubenswrapper[4808]: E1124 17:42:14.034142 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3563a459-e344-43c7-83ee-3da790cfe660" containerName="mariadb-account-create-update" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.034162 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="3563a459-e344-43c7-83ee-3da790cfe660" containerName="mariadb-account-create-update" Nov 24 17:42:14 crc kubenswrapper[4808]: E1124 17:42:14.034187 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c" containerName="mariadb-database-create" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.034198 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c" containerName="mariadb-database-create" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.034363 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c" containerName="mariadb-database-create" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.034378 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="3563a459-e344-43c7-83ee-3da790cfe660" containerName="mariadb-account-create-update" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.034958 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fmsjw" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.035701 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd9gc\" (UniqueName: \"kubernetes.io/projected/3563a459-e344-43c7-83ee-3da790cfe660-kube-api-access-wd9gc\") pod \"3563a459-e344-43c7-83ee-3da790cfe660\" (UID: \"3563a459-e344-43c7-83ee-3da790cfe660\") " Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.035869 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3563a459-e344-43c7-83ee-3da790cfe660-operator-scripts\") pod \"3563a459-e344-43c7-83ee-3da790cfe660\" (UID: \"3563a459-e344-43c7-83ee-3da790cfe660\") " Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.035941 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gr26w\" (UniqueName: \"kubernetes.io/projected/86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c-kube-api-access-gr26w\") pod \"86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c\" (UID: \"86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c\") " Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.036046 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c-operator-scripts\") pod \"86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c\" (UID: \"86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c\") " Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.036394 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fclf\" (UniqueName: \"kubernetes.io/projected/6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0-kube-api-access-2fclf\") pod \"keystone-5213-account-create-update-q7rbp\" (UID: \"6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0\") " pod="openstack/keystone-5213-account-create-update-q7rbp" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.036438 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0-operator-scripts\") pod \"keystone-5213-account-create-update-q7rbp\" (UID: \"6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0\") " pod="openstack/keystone-5213-account-create-update-q7rbp" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.036695 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c" (UID: "86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.036728 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3563a459-e344-43c7-83ee-3da790cfe660-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3563a459-e344-43c7-83ee-3da790cfe660" (UID: "3563a459-e344-43c7-83ee-3da790cfe660"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.037596 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0-operator-scripts\") pod \"keystone-5213-account-create-update-q7rbp\" (UID: \"6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0\") " pod="openstack/keystone-5213-account-create-update-q7rbp" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.040457 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3563a459-e344-43c7-83ee-3da790cfe660-kube-api-access-wd9gc" (OuterVolumeSpecName: "kube-api-access-wd9gc") pod "3563a459-e344-43c7-83ee-3da790cfe660" (UID: "3563a459-e344-43c7-83ee-3da790cfe660"). InnerVolumeSpecName "kube-api-access-wd9gc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.041148 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c-kube-api-access-gr26w" (OuterVolumeSpecName: "kube-api-access-gr26w") pod "86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c" (UID: "86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c"). InnerVolumeSpecName "kube-api-access-gr26w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.050675 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-fmsjw"] Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.058227 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fclf\" (UniqueName: \"kubernetes.io/projected/6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0-kube-api-access-2fclf\") pod \"keystone-5213-account-create-update-q7rbp\" (UID: \"6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0\") " pod="openstack/keystone-5213-account-create-update-q7rbp" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.137704 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39438be0-89df-4dff-98f3-368cb216fb6c-operator-scripts\") pod \"placement-db-create-fmsjw\" (UID: \"39438be0-89df-4dff-98f3-368cb216fb6c\") " pod="openstack/placement-db-create-fmsjw" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.137764 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf48r\" (UniqueName: \"kubernetes.io/projected/39438be0-89df-4dff-98f3-368cb216fb6c-kube-api-access-jf48r\") pod \"placement-db-create-fmsjw\" (UID: \"39438be0-89df-4dff-98f3-368cb216fb6c\") " pod="openstack/placement-db-create-fmsjw" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.137813 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3563a459-e344-43c7-83ee-3da790cfe660-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.137825 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gr26w\" (UniqueName: \"kubernetes.io/projected/86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c-kube-api-access-gr26w\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.137835 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.137859 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd9gc\" (UniqueName: \"kubernetes.io/projected/3563a459-e344-43c7-83ee-3da790cfe660-kube-api-access-wd9gc\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.139788 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6ab9-account-create-update-x9n7t"] Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.140876 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6ab9-account-create-update-x9n7t" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.145438 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.147945 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6ab9-account-create-update-x9n7t"] Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.183356 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9jp7n" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.209876 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5213-account-create-update-q7rbp" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.238891 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwtgp\" (UniqueName: \"kubernetes.io/projected/7c2587b0-bdda-4877-8847-c8af2afb7735-kube-api-access-qwtgp\") pod \"placement-6ab9-account-create-update-x9n7t\" (UID: \"7c2587b0-bdda-4877-8847-c8af2afb7735\") " pod="openstack/placement-6ab9-account-create-update-x9n7t" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.238984 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39438be0-89df-4dff-98f3-368cb216fb6c-operator-scripts\") pod \"placement-db-create-fmsjw\" (UID: \"39438be0-89df-4dff-98f3-368cb216fb6c\") " pod="openstack/placement-db-create-fmsjw" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.239059 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf48r\" (UniqueName: \"kubernetes.io/projected/39438be0-89df-4dff-98f3-368cb216fb6c-kube-api-access-jf48r\") pod \"placement-db-create-fmsjw\" (UID: \"39438be0-89df-4dff-98f3-368cb216fb6c\") " pod="openstack/placement-db-create-fmsjw" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.239134 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c2587b0-bdda-4877-8847-c8af2afb7735-operator-scripts\") pod \"placement-6ab9-account-create-update-x9n7t\" (UID: \"7c2587b0-bdda-4877-8847-c8af2afb7735\") " pod="openstack/placement-6ab9-account-create-update-x9n7t" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.239994 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39438be0-89df-4dff-98f3-368cb216fb6c-operator-scripts\") pod \"placement-db-create-fmsjw\" (UID: \"39438be0-89df-4dff-98f3-368cb216fb6c\") " pod="openstack/placement-db-create-fmsjw" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.259163 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf48r\" (UniqueName: \"kubernetes.io/projected/39438be0-89df-4dff-98f3-368cb216fb6c-kube-api-access-jf48r\") pod \"placement-db-create-fmsjw\" (UID: \"39438be0-89df-4dff-98f3-368cb216fb6c\") " pod="openstack/placement-db-create-fmsjw" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.343967 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c2587b0-bdda-4877-8847-c8af2afb7735-operator-scripts\") pod \"placement-6ab9-account-create-update-x9n7t\" (UID: \"7c2587b0-bdda-4877-8847-c8af2afb7735\") " pod="openstack/placement-6ab9-account-create-update-x9n7t" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.344179 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwtgp\" (UniqueName: \"kubernetes.io/projected/7c2587b0-bdda-4877-8847-c8af2afb7735-kube-api-access-qwtgp\") pod \"placement-6ab9-account-create-update-x9n7t\" (UID: \"7c2587b0-bdda-4877-8847-c8af2afb7735\") " pod="openstack/placement-6ab9-account-create-update-x9n7t" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.344831 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c2587b0-bdda-4877-8847-c8af2afb7735-operator-scripts\") pod \"placement-6ab9-account-create-update-x9n7t\" (UID: \"7c2587b0-bdda-4877-8847-c8af2afb7735\") " pod="openstack/placement-6ab9-account-create-update-x9n7t" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.364039 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwtgp\" (UniqueName: \"kubernetes.io/projected/7c2587b0-bdda-4877-8847-c8af2afb7735-kube-api-access-qwtgp\") pod \"placement-6ab9-account-create-update-x9n7t\" (UID: \"7c2587b0-bdda-4877-8847-c8af2afb7735\") " pod="openstack/placement-6ab9-account-create-update-x9n7t" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.410317 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fmsjw" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.454357 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8z24k" event={"ID":"86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c","Type":"ContainerDied","Data":"786a5ad7ccd68a05d270d505b7682f0d984375f225e42be77c919f0efea58f62"} Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.454396 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="786a5ad7ccd68a05d270d505b7682f0d984375f225e42be77c919f0efea58f62" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.454457 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8z24k" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.457559 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b750-account-create-update-v7jcb" event={"ID":"3563a459-e344-43c7-83ee-3da790cfe660","Type":"ContainerDied","Data":"58d6363d4225f7e1beb5ead6ddfce611c3e5b3d05cd1b5887f1827dd5d4dfdd6"} Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.457591 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58d6363d4225f7e1beb5ead6ddfce611c3e5b3d05cd1b5887f1827dd5d4dfdd6" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.457638 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b750-account-create-update-v7jcb" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.457641 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6ab9-account-create-update-x9n7t" Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.618959 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9jp7n"] Nov 24 17:42:14 crc kubenswrapper[4808]: W1124 17:42:14.625352 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef8985b9_2261_40cf_bfa9_ef86fb447616.slice/crio-e40768e6b4ae6d77d9b79b5c493801f108945b951e449d898d48186dccb02338 WatchSource:0}: Error finding container e40768e6b4ae6d77d9b79b5c493801f108945b951e449d898d48186dccb02338: Status 404 returned error can't find the container with id e40768e6b4ae6d77d9b79b5c493801f108945b951e449d898d48186dccb02338 Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.701848 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5213-account-create-update-q7rbp"] Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.859760 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-fmsjw"] Nov 24 17:42:14 crc kubenswrapper[4808]: W1124 17:42:14.865818 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39438be0_89df_4dff_98f3_368cb216fb6c.slice/crio-e2d1b133f6573032243642c7b94ee87e88ed9c8580addb691389c28b3afdc116 WatchSource:0}: Error finding container e2d1b133f6573032243642c7b94ee87e88ed9c8580addb691389c28b3afdc116: Status 404 returned error can't find the container with id e2d1b133f6573032243642c7b94ee87e88ed9c8580addb691389c28b3afdc116 Nov 24 17:42:14 crc kubenswrapper[4808]: I1124 17:42:14.958387 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6ab9-account-create-update-x9n7t"] Nov 24 17:42:15 crc kubenswrapper[4808]: I1124 17:42:15.056716 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:15 crc kubenswrapper[4808]: E1124 17:42:15.056998 4808 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 17:42:15 crc kubenswrapper[4808]: E1124 17:42:15.057061 4808 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 17:42:15 crc kubenswrapper[4808]: E1124 17:42:15.057128 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift podName:c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14 nodeName:}" failed. No retries permitted until 2025-11-24 17:42:23.057107646 +0000 UTC m=+935.654775438 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift") pod "swift-storage-0" (UID: "c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14") : configmap "swift-ring-files" not found Nov 24 17:42:15 crc kubenswrapper[4808]: I1124 17:42:15.468865 4808 generic.go:334] "Generic (PLEG): container finished" podID="ef8985b9-2261-40cf-bfa9-ef86fb447616" containerID="0581e349c664de7c6afa7e52742d3c29dea51433c46495c3e6012ba9da3c60e4" exitCode=0 Nov 24 17:42:15 crc kubenswrapper[4808]: I1124 17:42:15.468956 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9jp7n" event={"ID":"ef8985b9-2261-40cf-bfa9-ef86fb447616","Type":"ContainerDied","Data":"0581e349c664de7c6afa7e52742d3c29dea51433c46495c3e6012ba9da3c60e4"} Nov 24 17:42:15 crc kubenswrapper[4808]: I1124 17:42:15.468994 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9jp7n" event={"ID":"ef8985b9-2261-40cf-bfa9-ef86fb447616","Type":"ContainerStarted","Data":"e40768e6b4ae6d77d9b79b5c493801f108945b951e449d898d48186dccb02338"} Nov 24 17:42:15 crc kubenswrapper[4808]: I1124 17:42:15.471396 4808 generic.go:334] "Generic (PLEG): container finished" podID="6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0" containerID="8df39cb91c81ad1cd4d30c723b62d14131e8133471312802206d869b1a6602d6" exitCode=0 Nov 24 17:42:15 crc kubenswrapper[4808]: I1124 17:42:15.471452 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5213-account-create-update-q7rbp" event={"ID":"6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0","Type":"ContainerDied","Data":"8df39cb91c81ad1cd4d30c723b62d14131e8133471312802206d869b1a6602d6"} Nov 24 17:42:15 crc kubenswrapper[4808]: I1124 17:42:15.471472 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5213-account-create-update-q7rbp" event={"ID":"6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0","Type":"ContainerStarted","Data":"06c571431dfa796e8648698dc89e3b38566acf34ade91e326e38defd6be5f5f0"} Nov 24 17:42:15 crc kubenswrapper[4808]: I1124 17:42:15.472757 4808 generic.go:334] "Generic (PLEG): container finished" podID="39438be0-89df-4dff-98f3-368cb216fb6c" containerID="20fc52ba1f354e04df85fd49f44578b343421e2d26d9e773d0051d5d4f845b1e" exitCode=0 Nov 24 17:42:15 crc kubenswrapper[4808]: I1124 17:42:15.472803 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-fmsjw" event={"ID":"39438be0-89df-4dff-98f3-368cb216fb6c","Type":"ContainerDied","Data":"20fc52ba1f354e04df85fd49f44578b343421e2d26d9e773d0051d5d4f845b1e"} Nov 24 17:42:15 crc kubenswrapper[4808]: I1124 17:42:15.472818 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-fmsjw" event={"ID":"39438be0-89df-4dff-98f3-368cb216fb6c","Type":"ContainerStarted","Data":"e2d1b133f6573032243642c7b94ee87e88ed9c8580addb691389c28b3afdc116"} Nov 24 17:42:15 crc kubenswrapper[4808]: I1124 17:42:15.473763 4808 generic.go:334] "Generic (PLEG): container finished" podID="7c2587b0-bdda-4877-8847-c8af2afb7735" containerID="77f633b41f6fa0b20850e9facbfc52c1e7860f9a15833d8b38b86789c1e895e1" exitCode=0 Nov 24 17:42:15 crc kubenswrapper[4808]: I1124 17:42:15.473789 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6ab9-account-create-update-x9n7t" event={"ID":"7c2587b0-bdda-4877-8847-c8af2afb7735","Type":"ContainerDied","Data":"77f633b41f6fa0b20850e9facbfc52c1e7860f9a15833d8b38b86789c1e895e1"} Nov 24 17:42:15 crc kubenswrapper[4808]: I1124 17:42:15.473825 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6ab9-account-create-update-x9n7t" event={"ID":"7c2587b0-bdda-4877-8847-c8af2afb7735","Type":"ContainerStarted","Data":"794244dd39acfbb633d69ee7f8f2835ba5cadba5ddbda5d655bd6fac27ba7668"} Nov 24 17:42:16 crc kubenswrapper[4808]: I1124 17:42:16.430261 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:16 crc kubenswrapper[4808]: I1124 17:42:16.494171 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-jw8jm"] Nov 24 17:42:16 crc kubenswrapper[4808]: I1124 17:42:16.496130 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" podUID="3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8" containerName="dnsmasq-dns" containerID="cri-o://4b0b2422b70954cbb976ef8a394f08d8782af6e91419908a53810b9673ffdd31" gracePeriod=10 Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.015456 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5213-account-create-update-q7rbp" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.134936 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0-operator-scripts\") pod \"6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0\" (UID: \"6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0\") " Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.135136 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fclf\" (UniqueName: \"kubernetes.io/projected/6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0-kube-api-access-2fclf\") pod \"6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0\" (UID: \"6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0\") " Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.138441 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0" (UID: "6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.142054 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0-kube-api-access-2fclf" (OuterVolumeSpecName: "kube-api-access-2fclf") pod "6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0" (UID: "6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0"). InnerVolumeSpecName "kube-api-access-2fclf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.238278 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.238334 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fclf\" (UniqueName: \"kubernetes.io/projected/6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0-kube-api-access-2fclf\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.271904 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fmsjw" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.286766 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9jp7n" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.300089 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6ab9-account-create-update-x9n7t" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.328395 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.442006 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-dns-svc\") pod \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.442116 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39438be0-89df-4dff-98f3-368cb216fb6c-operator-scripts\") pod \"39438be0-89df-4dff-98f3-368cb216fb6c\" (UID: \"39438be0-89df-4dff-98f3-368cb216fb6c\") " Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.442189 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c2587b0-bdda-4877-8847-c8af2afb7735-operator-scripts\") pod \"7c2587b0-bdda-4877-8847-c8af2afb7735\" (UID: \"7c2587b0-bdda-4877-8847-c8af2afb7735\") " Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.442231 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxqzs\" (UniqueName: \"kubernetes.io/projected/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-kube-api-access-jxqzs\") pod \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.442302 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-ovsdbserver-sb\") pod \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.442325 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwtgp\" (UniqueName: \"kubernetes.io/projected/7c2587b0-bdda-4877-8847-c8af2afb7735-kube-api-access-qwtgp\") pod \"7c2587b0-bdda-4877-8847-c8af2afb7735\" (UID: \"7c2587b0-bdda-4877-8847-c8af2afb7735\") " Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.442344 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sn4l\" (UniqueName: \"kubernetes.io/projected/ef8985b9-2261-40cf-bfa9-ef86fb447616-kube-api-access-2sn4l\") pod \"ef8985b9-2261-40cf-bfa9-ef86fb447616\" (UID: \"ef8985b9-2261-40cf-bfa9-ef86fb447616\") " Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.442390 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-ovsdbserver-nb\") pod \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.442446 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-config\") pod \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\" (UID: \"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8\") " Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.442473 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jf48r\" (UniqueName: \"kubernetes.io/projected/39438be0-89df-4dff-98f3-368cb216fb6c-kube-api-access-jf48r\") pod \"39438be0-89df-4dff-98f3-368cb216fb6c\" (UID: \"39438be0-89df-4dff-98f3-368cb216fb6c\") " Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.442504 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef8985b9-2261-40cf-bfa9-ef86fb447616-operator-scripts\") pod \"ef8985b9-2261-40cf-bfa9-ef86fb447616\" (UID: \"ef8985b9-2261-40cf-bfa9-ef86fb447616\") " Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.443466 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef8985b9-2261-40cf-bfa9-ef86fb447616-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ef8985b9-2261-40cf-bfa9-ef86fb447616" (UID: "ef8985b9-2261-40cf-bfa9-ef86fb447616"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.444003 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39438be0-89df-4dff-98f3-368cb216fb6c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "39438be0-89df-4dff-98f3-368cb216fb6c" (UID: "39438be0-89df-4dff-98f3-368cb216fb6c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.444372 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c2587b0-bdda-4877-8847-c8af2afb7735-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7c2587b0-bdda-4877-8847-c8af2afb7735" (UID: "7c2587b0-bdda-4877-8847-c8af2afb7735"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.449495 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef8985b9-2261-40cf-bfa9-ef86fb447616-kube-api-access-2sn4l" (OuterVolumeSpecName: "kube-api-access-2sn4l") pod "ef8985b9-2261-40cf-bfa9-ef86fb447616" (UID: "ef8985b9-2261-40cf-bfa9-ef86fb447616"). InnerVolumeSpecName "kube-api-access-2sn4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.449636 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-kube-api-access-jxqzs" (OuterVolumeSpecName: "kube-api-access-jxqzs") pod "3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8" (UID: "3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8"). InnerVolumeSpecName "kube-api-access-jxqzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.452703 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c2587b0-bdda-4877-8847-c8af2afb7735-kube-api-access-qwtgp" (OuterVolumeSpecName: "kube-api-access-qwtgp") pod "7c2587b0-bdda-4877-8847-c8af2afb7735" (UID: "7c2587b0-bdda-4877-8847-c8af2afb7735"). InnerVolumeSpecName "kube-api-access-qwtgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.452903 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39438be0-89df-4dff-98f3-368cb216fb6c-kube-api-access-jf48r" (OuterVolumeSpecName: "kube-api-access-jf48r") pod "39438be0-89df-4dff-98f3-368cb216fb6c" (UID: "39438be0-89df-4dff-98f3-368cb216fb6c"). InnerVolumeSpecName "kube-api-access-jf48r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.493934 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9jp7n" event={"ID":"ef8985b9-2261-40cf-bfa9-ef86fb447616","Type":"ContainerDied","Data":"e40768e6b4ae6d77d9b79b5c493801f108945b951e449d898d48186dccb02338"} Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.493978 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9jp7n" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.493991 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e40768e6b4ae6d77d9b79b5c493801f108945b951e449d898d48186dccb02338" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.496817 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8" (UID: "3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.497177 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5213-account-create-update-q7rbp" event={"ID":"6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0","Type":"ContainerDied","Data":"06c571431dfa796e8648698dc89e3b38566acf34ade91e326e38defd6be5f5f0"} Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.497299 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06c571431dfa796e8648698dc89e3b38566acf34ade91e326e38defd6be5f5f0" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.497204 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5213-account-create-update-q7rbp" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.500208 4808 generic.go:334] "Generic (PLEG): container finished" podID="3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8" containerID="4b0b2422b70954cbb976ef8a394f08d8782af6e91419908a53810b9673ffdd31" exitCode=0 Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.500289 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" event={"ID":"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8","Type":"ContainerDied","Data":"4b0b2422b70954cbb976ef8a394f08d8782af6e91419908a53810b9673ffdd31"} Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.500322 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" event={"ID":"3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8","Type":"ContainerDied","Data":"115559c938bc8193bf85261ab8e505c009a259b406008ab2d58219625f5f41ea"} Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.500345 4808 scope.go:117] "RemoveContainer" containerID="4b0b2422b70954cbb976ef8a394f08d8782af6e91419908a53810b9673ffdd31" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.500530 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-jw8jm" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.501676 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-config" (OuterVolumeSpecName: "config") pod "3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8" (UID: "3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.504281 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-fmsjw" event={"ID":"39438be0-89df-4dff-98f3-368cb216fb6c","Type":"ContainerDied","Data":"e2d1b133f6573032243642c7b94ee87e88ed9c8580addb691389c28b3afdc116"} Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.504302 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2d1b133f6573032243642c7b94ee87e88ed9c8580addb691389c28b3afdc116" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.504343 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fmsjw" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.506633 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8" (UID: "3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.507374 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6ab9-account-create-update-x9n7t" event={"ID":"7c2587b0-bdda-4877-8847-c8af2afb7735","Type":"ContainerDied","Data":"794244dd39acfbb633d69ee7f8f2835ba5cadba5ddbda5d655bd6fac27ba7668"} Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.507441 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="794244dd39acfbb633d69ee7f8f2835ba5cadba5ddbda5d655bd6fac27ba7668" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.507633 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6ab9-account-create-update-x9n7t" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.515856 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8" (UID: "3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.528799 4808 scope.go:117] "RemoveContainer" containerID="bf5d8ef12838a98bbe4f640bfd36ad9f3220b4a6fdb3147e2a20a41d732c40f8" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.544560 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef8985b9-2261-40cf-bfa9-ef86fb447616-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.545261 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.545282 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39438be0-89df-4dff-98f3-368cb216fb6c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.545294 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c2587b0-bdda-4877-8847-c8af2afb7735-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.545307 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxqzs\" (UniqueName: \"kubernetes.io/projected/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-kube-api-access-jxqzs\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.545321 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.545335 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwtgp\" (UniqueName: \"kubernetes.io/projected/7c2587b0-bdda-4877-8847-c8af2afb7735-kube-api-access-qwtgp\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.545347 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sn4l\" (UniqueName: \"kubernetes.io/projected/ef8985b9-2261-40cf-bfa9-ef86fb447616-kube-api-access-2sn4l\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.545358 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.545370 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.545384 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jf48r\" (UniqueName: \"kubernetes.io/projected/39438be0-89df-4dff-98f3-368cb216fb6c-kube-api-access-jf48r\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.559464 4808 scope.go:117] "RemoveContainer" containerID="4b0b2422b70954cbb976ef8a394f08d8782af6e91419908a53810b9673ffdd31" Nov 24 17:42:17 crc kubenswrapper[4808]: E1124 17:42:17.560620 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b0b2422b70954cbb976ef8a394f08d8782af6e91419908a53810b9673ffdd31\": container with ID starting with 4b0b2422b70954cbb976ef8a394f08d8782af6e91419908a53810b9673ffdd31 not found: ID does not exist" containerID="4b0b2422b70954cbb976ef8a394f08d8782af6e91419908a53810b9673ffdd31" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.560693 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b0b2422b70954cbb976ef8a394f08d8782af6e91419908a53810b9673ffdd31"} err="failed to get container status \"4b0b2422b70954cbb976ef8a394f08d8782af6e91419908a53810b9673ffdd31\": rpc error: code = NotFound desc = could not find container \"4b0b2422b70954cbb976ef8a394f08d8782af6e91419908a53810b9673ffdd31\": container with ID starting with 4b0b2422b70954cbb976ef8a394f08d8782af6e91419908a53810b9673ffdd31 not found: ID does not exist" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.560741 4808 scope.go:117] "RemoveContainer" containerID="bf5d8ef12838a98bbe4f640bfd36ad9f3220b4a6fdb3147e2a20a41d732c40f8" Nov 24 17:42:17 crc kubenswrapper[4808]: E1124 17:42:17.561162 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf5d8ef12838a98bbe4f640bfd36ad9f3220b4a6fdb3147e2a20a41d732c40f8\": container with ID starting with bf5d8ef12838a98bbe4f640bfd36ad9f3220b4a6fdb3147e2a20a41d732c40f8 not found: ID does not exist" containerID="bf5d8ef12838a98bbe4f640bfd36ad9f3220b4a6fdb3147e2a20a41d732c40f8" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.561201 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf5d8ef12838a98bbe4f640bfd36ad9f3220b4a6fdb3147e2a20a41d732c40f8"} err="failed to get container status \"bf5d8ef12838a98bbe4f640bfd36ad9f3220b4a6fdb3147e2a20a41d732c40f8\": rpc error: code = NotFound desc = could not find container \"bf5d8ef12838a98bbe4f640bfd36ad9f3220b4a6fdb3147e2a20a41d732c40f8\": container with ID starting with bf5d8ef12838a98bbe4f640bfd36ad9f3220b4a6fdb3147e2a20a41d732c40f8 not found: ID does not exist" Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.829284 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-jw8jm"] Nov 24 17:42:17 crc kubenswrapper[4808]: I1124 17:42:17.836177 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-jw8jm"] Nov 24 17:42:18 crc kubenswrapper[4808]: I1124 17:42:18.362561 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8" path="/var/lib/kubelet/pods/3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8/volumes" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.519686 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-d6rlt"] Nov 24 17:42:19 crc kubenswrapper[4808]: E1124 17:42:19.520090 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39438be0-89df-4dff-98f3-368cb216fb6c" containerName="mariadb-database-create" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.520108 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="39438be0-89df-4dff-98f3-368cb216fb6c" containerName="mariadb-database-create" Nov 24 17:42:19 crc kubenswrapper[4808]: E1124 17:42:19.520128 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef8985b9-2261-40cf-bfa9-ef86fb447616" containerName="mariadb-database-create" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.520135 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef8985b9-2261-40cf-bfa9-ef86fb447616" containerName="mariadb-database-create" Nov 24 17:42:19 crc kubenswrapper[4808]: E1124 17:42:19.520164 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0" containerName="mariadb-account-create-update" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.520172 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0" containerName="mariadb-account-create-update" Nov 24 17:42:19 crc kubenswrapper[4808]: E1124 17:42:19.520186 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8" containerName="init" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.520194 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8" containerName="init" Nov 24 17:42:19 crc kubenswrapper[4808]: E1124 17:42:19.520204 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8" containerName="dnsmasq-dns" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.520211 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8" containerName="dnsmasq-dns" Nov 24 17:42:19 crc kubenswrapper[4808]: E1124 17:42:19.520222 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c2587b0-bdda-4877-8847-c8af2afb7735" containerName="mariadb-account-create-update" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.520229 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c2587b0-bdda-4877-8847-c8af2afb7735" containerName="mariadb-account-create-update" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.520413 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fbf2c5a-dd4f-46b2-a6b1-0871229cddf8" containerName="dnsmasq-dns" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.520430 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="39438be0-89df-4dff-98f3-368cb216fb6c" containerName="mariadb-database-create" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.520441 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef8985b9-2261-40cf-bfa9-ef86fb447616" containerName="mariadb-database-create" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.520456 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0" containerName="mariadb-account-create-update" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.520474 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c2587b0-bdda-4877-8847-c8af2afb7735" containerName="mariadb-account-create-update" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.521149 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.523972 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xzt98" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.525151 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.527489 4808 generic.go:334] "Generic (PLEG): container finished" podID="ee5479fc-5360-4541-990d-c8d1a97bae29" containerID="55772800d663664864db0a964b8e4d448e570041634e7cd558dc2ede4909964b" exitCode=0 Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.527528 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6ttz5" event={"ID":"ee5479fc-5360-4541-990d-c8d1a97bae29","Type":"ContainerDied","Data":"55772800d663664864db0a964b8e4d448e570041634e7cd558dc2ede4909964b"} Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.537858 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-d6rlt"] Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.681905 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-config-data\") pod \"glance-db-sync-d6rlt\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.682161 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjx82\" (UniqueName: \"kubernetes.io/projected/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-kube-api-access-sjx82\") pod \"glance-db-sync-d6rlt\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.682220 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-combined-ca-bundle\") pod \"glance-db-sync-d6rlt\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.682280 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-db-sync-config-data\") pod \"glance-db-sync-d6rlt\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.783603 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-config-data\") pod \"glance-db-sync-d6rlt\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.783693 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjx82\" (UniqueName: \"kubernetes.io/projected/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-kube-api-access-sjx82\") pod \"glance-db-sync-d6rlt\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.783715 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-combined-ca-bundle\") pod \"glance-db-sync-d6rlt\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.783740 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-db-sync-config-data\") pod \"glance-db-sync-d6rlt\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.790223 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-db-sync-config-data\") pod \"glance-db-sync-d6rlt\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.790499 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-combined-ca-bundle\") pod \"glance-db-sync-d6rlt\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.791694 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-config-data\") pod \"glance-db-sync-d6rlt\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.801674 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjx82\" (UniqueName: \"kubernetes.io/projected/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-kube-api-access-sjx82\") pod \"glance-db-sync-d6rlt\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:19 crc kubenswrapper[4808]: I1124 17:42:19.846145 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:20 crc kubenswrapper[4808]: I1124 17:42:20.011890 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 24 17:42:20 crc kubenswrapper[4808]: I1124 17:42:20.396885 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-d6rlt"] Nov 24 17:42:20 crc kubenswrapper[4808]: I1124 17:42:20.539627 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-d6rlt" event={"ID":"813f14a2-3982-4e6d-aca7-109d2bf2d1cb","Type":"ContainerStarted","Data":"7132ebd5e3473eb27bc395888ab3e16cacbbe47f0da17deef297aebad7b51824"} Nov 24 17:42:20 crc kubenswrapper[4808]: I1124 17:42:20.881530 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.003881 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq55s\" (UniqueName: \"kubernetes.io/projected/ee5479fc-5360-4541-990d-c8d1a97bae29-kube-api-access-jq55s\") pod \"ee5479fc-5360-4541-990d-c8d1a97bae29\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.004043 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ee5479fc-5360-4541-990d-c8d1a97bae29-etc-swift\") pod \"ee5479fc-5360-4541-990d-c8d1a97bae29\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.004082 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ee5479fc-5360-4541-990d-c8d1a97bae29-ring-data-devices\") pod \"ee5479fc-5360-4541-990d-c8d1a97bae29\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.004137 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-swiftconf\") pod \"ee5479fc-5360-4541-990d-c8d1a97bae29\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.004159 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee5479fc-5360-4541-990d-c8d1a97bae29-scripts\") pod \"ee5479fc-5360-4541-990d-c8d1a97bae29\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.004193 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-dispersionconf\") pod \"ee5479fc-5360-4541-990d-c8d1a97bae29\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.004246 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-combined-ca-bundle\") pod \"ee5479fc-5360-4541-990d-c8d1a97bae29\" (UID: \"ee5479fc-5360-4541-990d-c8d1a97bae29\") " Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.005493 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee5479fc-5360-4541-990d-c8d1a97bae29-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ee5479fc-5360-4541-990d-c8d1a97bae29" (UID: "ee5479fc-5360-4541-990d-c8d1a97bae29"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.005720 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee5479fc-5360-4541-990d-c8d1a97bae29-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ee5479fc-5360-4541-990d-c8d1a97bae29" (UID: "ee5479fc-5360-4541-990d-c8d1a97bae29"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.012123 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee5479fc-5360-4541-990d-c8d1a97bae29-kube-api-access-jq55s" (OuterVolumeSpecName: "kube-api-access-jq55s") pod "ee5479fc-5360-4541-990d-c8d1a97bae29" (UID: "ee5479fc-5360-4541-990d-c8d1a97bae29"). InnerVolumeSpecName "kube-api-access-jq55s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.014914 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ee5479fc-5360-4541-990d-c8d1a97bae29" (UID: "ee5479fc-5360-4541-990d-c8d1a97bae29"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.026787 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee5479fc-5360-4541-990d-c8d1a97bae29-scripts" (OuterVolumeSpecName: "scripts") pod "ee5479fc-5360-4541-990d-c8d1a97bae29" (UID: "ee5479fc-5360-4541-990d-c8d1a97bae29"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.030071 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee5479fc-5360-4541-990d-c8d1a97bae29" (UID: "ee5479fc-5360-4541-990d-c8d1a97bae29"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.034223 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ee5479fc-5360-4541-990d-c8d1a97bae29" (UID: "ee5479fc-5360-4541-990d-c8d1a97bae29"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.106090 4808 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ee5479fc-5360-4541-990d-c8d1a97bae29-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.106147 4808 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.106167 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee5479fc-5360-4541-990d-c8d1a97bae29-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.106185 4808 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.106206 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee5479fc-5360-4541-990d-c8d1a97bae29-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.106224 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq55s\" (UniqueName: \"kubernetes.io/projected/ee5479fc-5360-4541-990d-c8d1a97bae29-kube-api-access-jq55s\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.106244 4808 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ee5479fc-5360-4541-990d-c8d1a97bae29-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.555332 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6ttz5" event={"ID":"ee5479fc-5360-4541-990d-c8d1a97bae29","Type":"ContainerDied","Data":"737eb8a706d08a39baddaded640e6d83dd3f6748855784baf344a5624ca18d00"} Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.555712 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="737eb8a706d08a39baddaded640e6d83dd3f6748855784baf344a5624ca18d00" Nov 24 17:42:21 crc kubenswrapper[4808]: I1124 17:42:21.555439 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6ttz5" Nov 24 17:42:23 crc kubenswrapper[4808]: I1124 17:42:23.138367 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:23 crc kubenswrapper[4808]: I1124 17:42:23.147123 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14-etc-swift\") pod \"swift-storage-0\" (UID: \"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14\") " pod="openstack/swift-storage-0" Nov 24 17:42:23 crc kubenswrapper[4808]: I1124 17:42:23.153647 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 17:42:23 crc kubenswrapper[4808]: I1124 17:42:23.723877 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 17:42:23 crc kubenswrapper[4808]: W1124 17:42:23.731148 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6e7dc0a_09ef_49f9_ae1d_b4bff6dfbb14.slice/crio-b5c5d777249aa7785fc18ec6c3de0bd2fc523e2279ab6a060b0e71c30b873714 WatchSource:0}: Error finding container b5c5d777249aa7785fc18ec6c3de0bd2fc523e2279ab6a060b0e71c30b873714: Status 404 returned error can't find the container with id b5c5d777249aa7785fc18ec6c3de0bd2fc523e2279ab6a060b0e71c30b873714 Nov 24 17:42:24 crc kubenswrapper[4808]: I1124 17:42:24.587841 4808 generic.go:334] "Generic (PLEG): container finished" podID="e0b9711c-4309-40aa-8902-e2ad54e47000" containerID="417e38f4fcc06e5f31f575f1e377136dd887c214621183abc1e021be79950f22" exitCode=0 Nov 24 17:42:24 crc kubenswrapper[4808]: I1124 17:42:24.587952 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e0b9711c-4309-40aa-8902-e2ad54e47000","Type":"ContainerDied","Data":"417e38f4fcc06e5f31f575f1e377136dd887c214621183abc1e021be79950f22"} Nov 24 17:42:24 crc kubenswrapper[4808]: I1124 17:42:24.593730 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"b5c5d777249aa7785fc18ec6c3de0bd2fc523e2279ab6a060b0e71c30b873714"} Nov 24 17:42:24 crc kubenswrapper[4808]: I1124 17:42:24.596347 4808 generic.go:334] "Generic (PLEG): container finished" podID="c03d629c-1abf-440f-b3c8-d70e93e36eb1" containerID="c746e5e9913fa0751076b716717cda0a4d7faa73d4ccd89866e3c4908aef04bf" exitCode=0 Nov 24 17:42:24 crc kubenswrapper[4808]: I1124 17:42:24.596424 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c03d629c-1abf-440f-b3c8-d70e93e36eb1","Type":"ContainerDied","Data":"c746e5e9913fa0751076b716717cda0a4d7faa73d4ccd89866e3c4908aef04bf"} Nov 24 17:42:24 crc kubenswrapper[4808]: I1124 17:42:24.703703 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-zrsgl" podUID="dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec" containerName="ovn-controller" probeResult="failure" output=< Nov 24 17:42:24 crc kubenswrapper[4808]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 17:42:24 crc kubenswrapper[4808]: > Nov 24 17:42:24 crc kubenswrapper[4808]: I1124 17:42:24.727826 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:42:24 crc kubenswrapper[4808]: I1124 17:42:24.727887 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-58sfc" Nov 24 17:42:24 crc kubenswrapper[4808]: I1124 17:42:24.932449 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zrsgl-config-vpbq5"] Nov 24 17:42:24 crc kubenswrapper[4808]: E1124 17:42:24.932835 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee5479fc-5360-4541-990d-c8d1a97bae29" containerName="swift-ring-rebalance" Nov 24 17:42:24 crc kubenswrapper[4808]: I1124 17:42:24.932853 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee5479fc-5360-4541-990d-c8d1a97bae29" containerName="swift-ring-rebalance" Nov 24 17:42:24 crc kubenswrapper[4808]: I1124 17:42:24.933057 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee5479fc-5360-4541-990d-c8d1a97bae29" containerName="swift-ring-rebalance" Nov 24 17:42:24 crc kubenswrapper[4808]: I1124 17:42:24.933557 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:24 crc kubenswrapper[4808]: I1124 17:42:24.944901 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 17:42:24 crc kubenswrapper[4808]: I1124 17:42:24.951485 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zrsgl-config-vpbq5"] Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.071216 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d6a808ff-483c-41f6-8cdc-47dc08c016f6-additional-scripts\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.071585 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrcbz\" (UniqueName: \"kubernetes.io/projected/d6a808ff-483c-41f6-8cdc-47dc08c016f6-kube-api-access-jrcbz\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.071640 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-run-ovn\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.071681 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-log-ovn\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.071741 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6a808ff-483c-41f6-8cdc-47dc08c016f6-scripts\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.071788 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-run\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.173806 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-run\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.173889 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d6a808ff-483c-41f6-8cdc-47dc08c016f6-additional-scripts\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.173957 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrcbz\" (UniqueName: \"kubernetes.io/projected/d6a808ff-483c-41f6-8cdc-47dc08c016f6-kube-api-access-jrcbz\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.174004 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-run-ovn\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.174049 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-log-ovn\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.174099 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6a808ff-483c-41f6-8cdc-47dc08c016f6-scripts\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.174979 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-run\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.175124 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-run-ovn\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.175192 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-log-ovn\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.175843 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d6a808ff-483c-41f6-8cdc-47dc08c016f6-additional-scripts\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.176038 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6a808ff-483c-41f6-8cdc-47dc08c016f6-scripts\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.197549 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrcbz\" (UniqueName: \"kubernetes.io/projected/d6a808ff-483c-41f6-8cdc-47dc08c016f6-kube-api-access-jrcbz\") pod \"ovn-controller-zrsgl-config-vpbq5\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.256540 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.641213 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"98442b8486a96df4b782918d6df00976a3af80fe1f558c382976b7bcf7162a27"} Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.641462 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"cf2689f429b4a8282ff1be0767521467047e4b4cf73faef90b137a278d432e2b"} Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.641472 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"a0be710cb695b6d29744eae56a38a61eaad492bbe2311c26de6715ad95d5e7ee"} Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.641479 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"ba3f83984327ae65fb9ab6b6b4688124244cfd3cec053c3bd9486d8e74a3c814"} Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.653575 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c03d629c-1abf-440f-b3c8-d70e93e36eb1","Type":"ContainerStarted","Data":"fc9eaa638279c561b3f3ba4fdf1622a85a985a4623528792cf646cc044e0b0c3"} Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.654816 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.668294 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e0b9711c-4309-40aa-8902-e2ad54e47000","Type":"ContainerStarted","Data":"34753fc59dc288446ec80afdcfddee939d496ccc71ebc89ccf085eeac368798c"} Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.668804 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.715436 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=49.2792927 podStartE2EDuration="56.715414692s" podCreationTimestamp="2025-11-24 17:41:29 +0000 UTC" firstStartedPulling="2025-11-24 17:41:42.605951651 +0000 UTC m=+895.203619453" lastFinishedPulling="2025-11-24 17:41:50.042073643 +0000 UTC m=+902.639741445" observedRunningTime="2025-11-24 17:42:25.713235589 +0000 UTC m=+938.310903391" watchObservedRunningTime="2025-11-24 17:42:25.715414692 +0000 UTC m=+938.313082494" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.741590 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=48.912592021 podStartE2EDuration="56.741572963s" podCreationTimestamp="2025-11-24 17:41:29 +0000 UTC" firstStartedPulling="2025-11-24 17:41:42.214481372 +0000 UTC m=+894.812149174" lastFinishedPulling="2025-11-24 17:41:50.043462314 +0000 UTC m=+902.641130116" observedRunningTime="2025-11-24 17:42:25.738541285 +0000 UTC m=+938.336209087" watchObservedRunningTime="2025-11-24 17:42:25.741572963 +0000 UTC m=+938.339240765" Nov 24 17:42:25 crc kubenswrapper[4808]: I1124 17:42:25.831403 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zrsgl-config-vpbq5"] Nov 24 17:42:26 crc kubenswrapper[4808]: I1124 17:42:26.678205 4808 generic.go:334] "Generic (PLEG): container finished" podID="d6a808ff-483c-41f6-8cdc-47dc08c016f6" containerID="d04912d83a47d1161dbc73f3b5b80134fa4d250a92b7071f0aef10f522bf0dbb" exitCode=0 Nov 24 17:42:26 crc kubenswrapper[4808]: I1124 17:42:26.678544 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zrsgl-config-vpbq5" event={"ID":"d6a808ff-483c-41f6-8cdc-47dc08c016f6","Type":"ContainerDied","Data":"d04912d83a47d1161dbc73f3b5b80134fa4d250a92b7071f0aef10f522bf0dbb"} Nov 24 17:42:26 crc kubenswrapper[4808]: I1124 17:42:26.679476 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zrsgl-config-vpbq5" event={"ID":"d6a808ff-483c-41f6-8cdc-47dc08c016f6","Type":"ContainerStarted","Data":"f2252815ba4319cd54ec250d08cf9d0188716891722d5ffe04f19295df10337a"} Nov 24 17:42:29 crc kubenswrapper[4808]: I1124 17:42:29.698231 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-zrsgl" Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.048457 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.206032 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6a808ff-483c-41f6-8cdc-47dc08c016f6-scripts\") pod \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.206143 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d6a808ff-483c-41f6-8cdc-47dc08c016f6-additional-scripts\") pod \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.206198 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrcbz\" (UniqueName: \"kubernetes.io/projected/d6a808ff-483c-41f6-8cdc-47dc08c016f6-kube-api-access-jrcbz\") pod \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.206219 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-log-ovn\") pod \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.206280 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-run-ovn\") pod \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.206303 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-run\") pod \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\" (UID: \"d6a808ff-483c-41f6-8cdc-47dc08c016f6\") " Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.206783 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-run" (OuterVolumeSpecName: "var-run") pod "d6a808ff-483c-41f6-8cdc-47dc08c016f6" (UID: "d6a808ff-483c-41f6-8cdc-47dc08c016f6"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.207643 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6a808ff-483c-41f6-8cdc-47dc08c016f6-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "d6a808ff-483c-41f6-8cdc-47dc08c016f6" (UID: "d6a808ff-483c-41f6-8cdc-47dc08c016f6"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.207907 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "d6a808ff-483c-41f6-8cdc-47dc08c016f6" (UID: "d6a808ff-483c-41f6-8cdc-47dc08c016f6"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.207947 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "d6a808ff-483c-41f6-8cdc-47dc08c016f6" (UID: "d6a808ff-483c-41f6-8cdc-47dc08c016f6"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.208526 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6a808ff-483c-41f6-8cdc-47dc08c016f6-scripts" (OuterVolumeSpecName: "scripts") pod "d6a808ff-483c-41f6-8cdc-47dc08c016f6" (UID: "d6a808ff-483c-41f6-8cdc-47dc08c016f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.222103 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6a808ff-483c-41f6-8cdc-47dc08c016f6-kube-api-access-jrcbz" (OuterVolumeSpecName: "kube-api-access-jrcbz") pod "d6a808ff-483c-41f6-8cdc-47dc08c016f6" (UID: "d6a808ff-483c-41f6-8cdc-47dc08c016f6"). InnerVolumeSpecName "kube-api-access-jrcbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.307954 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6a808ff-483c-41f6-8cdc-47dc08c016f6-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.308006 4808 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d6a808ff-483c-41f6-8cdc-47dc08c016f6-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.308042 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrcbz\" (UniqueName: \"kubernetes.io/projected/d6a808ff-483c-41f6-8cdc-47dc08c016f6-kube-api-access-jrcbz\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.308053 4808 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.308064 4808 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.308074 4808 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d6a808ff-483c-41f6-8cdc-47dc08c016f6-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.756200 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"8c4229bbdf3dd6df4dd49c8a11f455e16208e653393752888544a27451d7505a"} Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.756536 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"ddc3fe7aa1f26807f01ee24da4cd78a3269d33775dafee4894c288aa66762502"} Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.760524 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zrsgl-config-vpbq5" event={"ID":"d6a808ff-483c-41f6-8cdc-47dc08c016f6","Type":"ContainerDied","Data":"f2252815ba4319cd54ec250d08cf9d0188716891722d5ffe04f19295df10337a"} Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.760561 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2252815ba4319cd54ec250d08cf9d0188716891722d5ffe04f19295df10337a" Nov 24 17:42:33 crc kubenswrapper[4808]: I1124 17:42:33.760640 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrsgl-config-vpbq5" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.141227 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zrsgl-config-vpbq5"] Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.151143 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zrsgl-config-vpbq5"] Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.270591 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zrsgl-config-zxlvh"] Nov 24 17:42:34 crc kubenswrapper[4808]: E1124 17:42:34.271125 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a808ff-483c-41f6-8cdc-47dc08c016f6" containerName="ovn-config" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.271149 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a808ff-483c-41f6-8cdc-47dc08c016f6" containerName="ovn-config" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.271456 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a808ff-483c-41f6-8cdc-47dc08c016f6" containerName="ovn-config" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.272740 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.275555 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.286059 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zrsgl-config-zxlvh"] Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.363482 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6a808ff-483c-41f6-8cdc-47dc08c016f6" path="/var/lib/kubelet/pods/d6a808ff-483c-41f6-8cdc-47dc08c016f6/volumes" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.424040 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn8th\" (UniqueName: \"kubernetes.io/projected/76cded39-57d8-4e88-a054-f7c0fe2863d7-kube-api-access-cn8th\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.425225 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-log-ovn\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.425406 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-run-ovn\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.425531 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-run\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.425715 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76cded39-57d8-4e88-a054-f7c0fe2863d7-scripts\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.425801 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/76cded39-57d8-4e88-a054-f7c0fe2863d7-additional-scripts\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.526779 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn8th\" (UniqueName: \"kubernetes.io/projected/76cded39-57d8-4e88-a054-f7c0fe2863d7-kube-api-access-cn8th\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.527073 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-log-ovn\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.527185 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-run-ovn\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.527289 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-run\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.527391 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/76cded39-57d8-4e88-a054-f7c0fe2863d7-additional-scripts\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.527455 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76cded39-57d8-4e88-a054-f7c0fe2863d7-scripts\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.528041 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-run-ovn\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.528448 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-log-ovn\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.528494 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-run\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.528977 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/76cded39-57d8-4e88-a054-f7c0fe2863d7-additional-scripts\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.530126 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76cded39-57d8-4e88-a054-f7c0fe2863d7-scripts\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.547342 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn8th\" (UniqueName: \"kubernetes.io/projected/76cded39-57d8-4e88-a054-f7c0fe2863d7-kube-api-access-cn8th\") pod \"ovn-controller-zrsgl-config-zxlvh\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.595297 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.784122 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"0270a405010a9caf2d0c6d77bffb88999767ca47e52f377bae6aa6cb2cec8d74"} Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.784720 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"9a8b1278a56a7ea7052354d012c5162b7c2806636e59ab27a5475ab096ee2115"} Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.785574 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-d6rlt" event={"ID":"813f14a2-3982-4e6d-aca7-109d2bf2d1cb","Type":"ContainerStarted","Data":"df8384c369f4dc156269d9d1914456e7b26d80de43e0054abef15709835f7c66"} Nov 24 17:42:34 crc kubenswrapper[4808]: I1124 17:42:34.808151 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-d6rlt" podStartSLOduration=2.824605274 podStartE2EDuration="15.80812851s" podCreationTimestamp="2025-11-24 17:42:19 +0000 UTC" firstStartedPulling="2025-11-24 17:42:20.404253322 +0000 UTC m=+933.001921124" lastFinishedPulling="2025-11-24 17:42:33.387776558 +0000 UTC m=+945.985444360" observedRunningTime="2025-11-24 17:42:34.804550926 +0000 UTC m=+947.402218758" watchObservedRunningTime="2025-11-24 17:42:34.80812851 +0000 UTC m=+947.405796332" Nov 24 17:42:35 crc kubenswrapper[4808]: I1124 17:42:35.099744 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zrsgl-config-zxlvh"] Nov 24 17:42:35 crc kubenswrapper[4808]: I1124 17:42:35.797653 4808 generic.go:334] "Generic (PLEG): container finished" podID="76cded39-57d8-4e88-a054-f7c0fe2863d7" containerID="9b40e76510974e9bec17fedccedaa306a188f9d7a1f0451ec0d8474a6d752e1c" exitCode=0 Nov 24 17:42:35 crc kubenswrapper[4808]: I1124 17:42:35.797819 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zrsgl-config-zxlvh" event={"ID":"76cded39-57d8-4e88-a054-f7c0fe2863d7","Type":"ContainerDied","Data":"9b40e76510974e9bec17fedccedaa306a188f9d7a1f0451ec0d8474a6d752e1c"} Nov 24 17:42:35 crc kubenswrapper[4808]: I1124 17:42:35.797987 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zrsgl-config-zxlvh" event={"ID":"76cded39-57d8-4e88-a054-f7c0fe2863d7","Type":"ContainerStarted","Data":"db39b053d8d3081b6bc5dc6a3c34c6514280b38c5ff6b0757c2ac303035f32a6"} Nov 24 17:42:35 crc kubenswrapper[4808]: I1124 17:42:35.807563 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"01b197e20eeca33386ca042d904fdabccadcbe474603532c8a0f0d249f9ac674"} Nov 24 17:42:35 crc kubenswrapper[4808]: I1124 17:42:35.807618 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"08c3df8407d871ce6aedb11bf9b5dec4d4910deeab190c309174a9218ffdca29"} Nov 24 17:42:35 crc kubenswrapper[4808]: I1124 17:42:35.807632 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"46feee14a36760c84bb924a1465a480e79c538218213a61d8b1d2d1d1b7e5595"} Nov 24 17:42:36 crc kubenswrapper[4808]: I1124 17:42:36.827011 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"bfeb5953c39eea26e4cf4d54794957c82a7332b58ec8ef79c7c0a3d89baae867"} Nov 24 17:42:36 crc kubenswrapper[4808]: I1124 17:42:36.827589 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"da89f155f4dbce307ba8e52b3a8f90796eef441e72c997159d2650d8f16ac672"} Nov 24 17:42:36 crc kubenswrapper[4808]: I1124 17:42:36.827603 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"76ac379d8f64fd0c25912fa82eee1cebc763e4ed78e29b1996ec9192052ea9e3"} Nov 24 17:42:36 crc kubenswrapper[4808]: I1124 17:42:36.827625 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14","Type":"ContainerStarted","Data":"74a094202011c239d4be69b027ff58014380770ab2f04e55302da4c2f4c444b6"} Nov 24 17:42:36 crc kubenswrapper[4808]: I1124 17:42:36.865232 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=19.375677467 podStartE2EDuration="30.865216738s" podCreationTimestamp="2025-11-24 17:42:06 +0000 UTC" firstStartedPulling="2025-11-24 17:42:23.740071422 +0000 UTC m=+936.337739224" lastFinishedPulling="2025-11-24 17:42:35.229610693 +0000 UTC m=+947.827278495" observedRunningTime="2025-11-24 17:42:36.860998906 +0000 UTC m=+949.458666738" watchObservedRunningTime="2025-11-24 17:42:36.865216738 +0000 UTC m=+949.462884540" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.186425 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.280048 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-2rpjx"] Nov 24 17:42:37 crc kubenswrapper[4808]: E1124 17:42:37.280453 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76cded39-57d8-4e88-a054-f7c0fe2863d7" containerName="ovn-config" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.280478 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="76cded39-57d8-4e88-a054-f7c0fe2863d7" containerName="ovn-config" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.280705 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="76cded39-57d8-4e88-a054-f7c0fe2863d7" containerName="ovn-config" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.282253 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.284584 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.290833 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-2rpjx"] Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.305468 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-run\") pod \"76cded39-57d8-4e88-a054-f7c0fe2863d7\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.305563 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-run" (OuterVolumeSpecName: "var-run") pod "76cded39-57d8-4e88-a054-f7c0fe2863d7" (UID: "76cded39-57d8-4e88-a054-f7c0fe2863d7"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.305859 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76cded39-57d8-4e88-a054-f7c0fe2863d7-scripts\") pod \"76cded39-57d8-4e88-a054-f7c0fe2863d7\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.305987 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn8th\" (UniqueName: \"kubernetes.io/projected/76cded39-57d8-4e88-a054-f7c0fe2863d7-kube-api-access-cn8th\") pod \"76cded39-57d8-4e88-a054-f7c0fe2863d7\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.306057 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-run-ovn\") pod \"76cded39-57d8-4e88-a054-f7c0fe2863d7\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.306096 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-log-ovn\") pod \"76cded39-57d8-4e88-a054-f7c0fe2863d7\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.306159 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/76cded39-57d8-4e88-a054-f7c0fe2863d7-additional-scripts\") pod \"76cded39-57d8-4e88-a054-f7c0fe2863d7\" (UID: \"76cded39-57d8-4e88-a054-f7c0fe2863d7\") " Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.306209 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "76cded39-57d8-4e88-a054-f7c0fe2863d7" (UID: "76cded39-57d8-4e88-a054-f7c0fe2863d7"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.306265 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "76cded39-57d8-4e88-a054-f7c0fe2863d7" (UID: "76cded39-57d8-4e88-a054-f7c0fe2863d7"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.306540 4808 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.306563 4808 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.306575 4808 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/76cded39-57d8-4e88-a054-f7c0fe2863d7-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.307009 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76cded39-57d8-4e88-a054-f7c0fe2863d7-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "76cded39-57d8-4e88-a054-f7c0fe2863d7" (UID: "76cded39-57d8-4e88-a054-f7c0fe2863d7"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.307329 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76cded39-57d8-4e88-a054-f7c0fe2863d7-scripts" (OuterVolumeSpecName: "scripts") pod "76cded39-57d8-4e88-a054-f7c0fe2863d7" (UID: "76cded39-57d8-4e88-a054-f7c0fe2863d7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.313370 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76cded39-57d8-4e88-a054-f7c0fe2863d7-kube-api-access-cn8th" (OuterVolumeSpecName: "kube-api-access-cn8th") pod "76cded39-57d8-4e88-a054-f7c0fe2863d7" (UID: "76cded39-57d8-4e88-a054-f7c0fe2863d7"). InnerVolumeSpecName "kube-api-access-cn8th". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.408379 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-config\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.408560 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b44qk\" (UniqueName: \"kubernetes.io/projected/4cd2504d-284c-47ef-8579-78a798d4a5cb-kube-api-access-b44qk\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.408623 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.408654 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-dns-svc\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.408770 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.408824 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.408939 4808 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/76cded39-57d8-4e88-a054-f7c0fe2863d7-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.408964 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76cded39-57d8-4e88-a054-f7c0fe2863d7-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.408977 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn8th\" (UniqueName: \"kubernetes.io/projected/76cded39-57d8-4e88-a054-f7c0fe2863d7-kube-api-access-cn8th\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.510749 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.510798 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-dns-svc\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.510837 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.510873 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.510952 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-config\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.510981 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b44qk\" (UniqueName: \"kubernetes.io/projected/4cd2504d-284c-47ef-8579-78a798d4a5cb-kube-api-access-b44qk\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.512146 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.512173 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-dns-svc\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.512205 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.512463 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-config\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.512877 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.526839 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b44qk\" (UniqueName: \"kubernetes.io/projected/4cd2504d-284c-47ef-8579-78a798d4a5cb-kube-api-access-b44qk\") pod \"dnsmasq-dns-764c5664d7-2rpjx\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.598635 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.846256 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrsgl-config-zxlvh" Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.849168 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zrsgl-config-zxlvh" event={"ID":"76cded39-57d8-4e88-a054-f7c0fe2863d7","Type":"ContainerDied","Data":"db39b053d8d3081b6bc5dc6a3c34c6514280b38c5ff6b0757c2ac303035f32a6"} Nov 24 17:42:37 crc kubenswrapper[4808]: I1124 17:42:37.849225 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db39b053d8d3081b6bc5dc6a3c34c6514280b38c5ff6b0757c2ac303035f32a6" Nov 24 17:42:38 crc kubenswrapper[4808]: I1124 17:42:38.123311 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-2rpjx"] Nov 24 17:42:38 crc kubenswrapper[4808]: W1124 17:42:38.131809 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4cd2504d_284c_47ef_8579_78a798d4a5cb.slice/crio-7f2999bf27f145294eda83e7d6802fe0cdae9e963e01ff95979c0623f8200927 WatchSource:0}: Error finding container 7f2999bf27f145294eda83e7d6802fe0cdae9e963e01ff95979c0623f8200927: Status 404 returned error can't find the container with id 7f2999bf27f145294eda83e7d6802fe0cdae9e963e01ff95979c0623f8200927 Nov 24 17:42:38 crc kubenswrapper[4808]: I1124 17:42:38.275178 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zrsgl-config-zxlvh"] Nov 24 17:42:38 crc kubenswrapper[4808]: I1124 17:42:38.284722 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zrsgl-config-zxlvh"] Nov 24 17:42:38 crc kubenswrapper[4808]: I1124 17:42:38.360710 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76cded39-57d8-4e88-a054-f7c0fe2863d7" path="/var/lib/kubelet/pods/76cded39-57d8-4e88-a054-f7c0fe2863d7/volumes" Nov 24 17:42:38 crc kubenswrapper[4808]: I1124 17:42:38.846702 4808 generic.go:334] "Generic (PLEG): container finished" podID="4cd2504d-284c-47ef-8579-78a798d4a5cb" containerID="6767976c7a11d87f388fa4e6cce062addc10ada3f5582a76073742371b53a0b3" exitCode=0 Nov 24 17:42:38 crc kubenswrapper[4808]: I1124 17:42:38.846740 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" event={"ID":"4cd2504d-284c-47ef-8579-78a798d4a5cb","Type":"ContainerDied","Data":"6767976c7a11d87f388fa4e6cce062addc10ada3f5582a76073742371b53a0b3"} Nov 24 17:42:38 crc kubenswrapper[4808]: I1124 17:42:38.846768 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" event={"ID":"4cd2504d-284c-47ef-8579-78a798d4a5cb","Type":"ContainerStarted","Data":"7f2999bf27f145294eda83e7d6802fe0cdae9e963e01ff95979c0623f8200927"} Nov 24 17:42:39 crc kubenswrapper[4808]: I1124 17:42:39.855895 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" event={"ID":"4cd2504d-284c-47ef-8579-78a798d4a5cb","Type":"ContainerStarted","Data":"69fae8344fc7fd25f56c916847ee0d2dabdfe0165e8919776a3be809d2b527af"} Nov 24 17:42:39 crc kubenswrapper[4808]: I1124 17:42:39.856221 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:39 crc kubenswrapper[4808]: I1124 17:42:39.873820 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" podStartSLOduration=2.873803258 podStartE2EDuration="2.873803258s" podCreationTimestamp="2025-11-24 17:42:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:42:39.87079145 +0000 UTC m=+952.468459263" watchObservedRunningTime="2025-11-24 17:42:39.873803258 +0000 UTC m=+952.471471060" Nov 24 17:42:40 crc kubenswrapper[4808]: I1124 17:42:40.661246 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 17:42:40 crc kubenswrapper[4808]: I1124 17:42:40.865397 4808 generic.go:334] "Generic (PLEG): container finished" podID="813f14a2-3982-4e6d-aca7-109d2bf2d1cb" containerID="df8384c369f4dc156269d9d1914456e7b26d80de43e0054abef15709835f7c66" exitCode=0 Nov 24 17:42:40 crc kubenswrapper[4808]: I1124 17:42:40.865506 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-d6rlt" event={"ID":"813f14a2-3982-4e6d-aca7-109d2bf2d1cb","Type":"ContainerDied","Data":"df8384c369f4dc156269d9d1914456e7b26d80de43e0054abef15709835f7c66"} Nov 24 17:42:40 crc kubenswrapper[4808]: I1124 17:42:40.937565 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-qg8hc"] Nov 24 17:42:40 crc kubenswrapper[4808]: I1124 17:42:40.938730 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qg8hc" Nov 24 17:42:40 crc kubenswrapper[4808]: I1124 17:42:40.949940 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qg8hc"] Nov 24 17:42:40 crc kubenswrapper[4808]: I1124 17:42:40.967436 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjzq4\" (UniqueName: \"kubernetes.io/projected/180dfd5f-4d34-42fc-983d-90a54e945a15-kube-api-access-bjzq4\") pod \"cinder-db-create-qg8hc\" (UID: \"180dfd5f-4d34-42fc-983d-90a54e945a15\") " pod="openstack/cinder-db-create-qg8hc" Nov 24 17:42:40 crc kubenswrapper[4808]: I1124 17:42:40.967598 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/180dfd5f-4d34-42fc-983d-90a54e945a15-operator-scripts\") pod \"cinder-db-create-qg8hc\" (UID: \"180dfd5f-4d34-42fc-983d-90a54e945a15\") " pod="openstack/cinder-db-create-qg8hc" Nov 24 17:42:40 crc kubenswrapper[4808]: I1124 17:42:40.970199 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.058679 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-b2sdj"] Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.059794 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b2sdj" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.069238 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjzq4\" (UniqueName: \"kubernetes.io/projected/180dfd5f-4d34-42fc-983d-90a54e945a15-kube-api-access-bjzq4\") pod \"cinder-db-create-qg8hc\" (UID: \"180dfd5f-4d34-42fc-983d-90a54e945a15\") " pod="openstack/cinder-db-create-qg8hc" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.069433 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/180dfd5f-4d34-42fc-983d-90a54e945a15-operator-scripts\") pod \"cinder-db-create-qg8hc\" (UID: \"180dfd5f-4d34-42fc-983d-90a54e945a15\") " pod="openstack/cinder-db-create-qg8hc" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.070374 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/180dfd5f-4d34-42fc-983d-90a54e945a15-operator-scripts\") pod \"cinder-db-create-qg8hc\" (UID: \"180dfd5f-4d34-42fc-983d-90a54e945a15\") " pod="openstack/cinder-db-create-qg8hc" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.079843 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-b2sdj"] Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.106514 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjzq4\" (UniqueName: \"kubernetes.io/projected/180dfd5f-4d34-42fc-983d-90a54e945a15-kube-api-access-bjzq4\") pod \"cinder-db-create-qg8hc\" (UID: \"180dfd5f-4d34-42fc-983d-90a54e945a15\") " pod="openstack/cinder-db-create-qg8hc" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.163196 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-cb38-account-create-update-zdknp"] Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.170231 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-cb38-account-create-update-zdknp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.170927 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9jv9\" (UniqueName: \"kubernetes.io/projected/04e46a77-a53c-4edd-a1d8-c74a7b61c666-kube-api-access-j9jv9\") pod \"barbican-db-create-b2sdj\" (UID: \"04e46a77-a53c-4edd-a1d8-c74a7b61c666\") " pod="openstack/barbican-db-create-b2sdj" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.170998 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04e46a77-a53c-4edd-a1d8-c74a7b61c666-operator-scripts\") pod \"barbican-db-create-b2sdj\" (UID: \"04e46a77-a53c-4edd-a1d8-c74a7b61c666\") " pod="openstack/barbican-db-create-b2sdj" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.173701 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.176705 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-cb38-account-create-update-zdknp"] Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.246870 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-df60-account-create-update-mlxn7"] Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.248144 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-df60-account-create-update-mlxn7" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.250302 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.253970 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qg8hc" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.262285 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-df60-account-create-update-mlxn7"] Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.273032 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9jv9\" (UniqueName: \"kubernetes.io/projected/04e46a77-a53c-4edd-a1d8-c74a7b61c666-kube-api-access-j9jv9\") pod \"barbican-db-create-b2sdj\" (UID: \"04e46a77-a53c-4edd-a1d8-c74a7b61c666\") " pod="openstack/barbican-db-create-b2sdj" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.273099 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv7kd\" (UniqueName: \"kubernetes.io/projected/f113a2b1-1597-48d7-9420-16d6f34fedf8-kube-api-access-vv7kd\") pod \"cinder-df60-account-create-update-mlxn7\" (UID: \"f113a2b1-1597-48d7-9420-16d6f34fedf8\") " pod="openstack/cinder-df60-account-create-update-mlxn7" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.273148 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f113a2b1-1597-48d7-9420-16d6f34fedf8-operator-scripts\") pod \"cinder-df60-account-create-update-mlxn7\" (UID: \"f113a2b1-1597-48d7-9420-16d6f34fedf8\") " pod="openstack/cinder-df60-account-create-update-mlxn7" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.273186 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04e46a77-a53c-4edd-a1d8-c74a7b61c666-operator-scripts\") pod \"barbican-db-create-b2sdj\" (UID: \"04e46a77-a53c-4edd-a1d8-c74a7b61c666\") " pod="openstack/barbican-db-create-b2sdj" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.273258 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9rfz\" (UniqueName: \"kubernetes.io/projected/53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a-kube-api-access-r9rfz\") pod \"barbican-cb38-account-create-update-zdknp\" (UID: \"53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a\") " pod="openstack/barbican-cb38-account-create-update-zdknp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.273322 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a-operator-scripts\") pod \"barbican-cb38-account-create-update-zdknp\" (UID: \"53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a\") " pod="openstack/barbican-cb38-account-create-update-zdknp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.273953 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04e46a77-a53c-4edd-a1d8-c74a7b61c666-operator-scripts\") pod \"barbican-db-create-b2sdj\" (UID: \"04e46a77-a53c-4edd-a1d8-c74a7b61c666\") " pod="openstack/barbican-db-create-b2sdj" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.305444 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9jv9\" (UniqueName: \"kubernetes.io/projected/04e46a77-a53c-4edd-a1d8-c74a7b61c666-kube-api-access-j9jv9\") pod \"barbican-db-create-b2sdj\" (UID: \"04e46a77-a53c-4edd-a1d8-c74a7b61c666\") " pod="openstack/barbican-db-create-b2sdj" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.374975 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9rfz\" (UniqueName: \"kubernetes.io/projected/53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a-kube-api-access-r9rfz\") pod \"barbican-cb38-account-create-update-zdknp\" (UID: \"53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a\") " pod="openstack/barbican-cb38-account-create-update-zdknp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.375073 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a-operator-scripts\") pod \"barbican-cb38-account-create-update-zdknp\" (UID: \"53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a\") " pod="openstack/barbican-cb38-account-create-update-zdknp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.375206 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv7kd\" (UniqueName: \"kubernetes.io/projected/f113a2b1-1597-48d7-9420-16d6f34fedf8-kube-api-access-vv7kd\") pod \"cinder-df60-account-create-update-mlxn7\" (UID: \"f113a2b1-1597-48d7-9420-16d6f34fedf8\") " pod="openstack/cinder-df60-account-create-update-mlxn7" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.375247 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f113a2b1-1597-48d7-9420-16d6f34fedf8-operator-scripts\") pod \"cinder-df60-account-create-update-mlxn7\" (UID: \"f113a2b1-1597-48d7-9420-16d6f34fedf8\") " pod="openstack/cinder-df60-account-create-update-mlxn7" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.377041 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a-operator-scripts\") pod \"barbican-cb38-account-create-update-zdknp\" (UID: \"53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a\") " pod="openstack/barbican-cb38-account-create-update-zdknp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.378677 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-5bhqp"] Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.380083 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-5bhqp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.381546 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b2sdj" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.383051 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f113a2b1-1597-48d7-9420-16d6f34fedf8-operator-scripts\") pod \"cinder-df60-account-create-update-mlxn7\" (UID: \"f113a2b1-1597-48d7-9420-16d6f34fedf8\") " pod="openstack/cinder-df60-account-create-update-mlxn7" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.390131 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-dr57t"] Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.395575 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.395837 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.395851 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.396354 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-622rb" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.399675 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9rfz\" (UniqueName: \"kubernetes.io/projected/53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a-kube-api-access-r9rfz\") pod \"barbican-cb38-account-create-update-zdknp\" (UID: \"53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a\") " pod="openstack/barbican-cb38-account-create-update-zdknp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.399961 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dr57t" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.404936 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-5bhqp"] Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.423333 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-dr57t"] Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.445105 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv7kd\" (UniqueName: \"kubernetes.io/projected/f113a2b1-1597-48d7-9420-16d6f34fedf8-kube-api-access-vv7kd\") pod \"cinder-df60-account-create-update-mlxn7\" (UID: \"f113a2b1-1597-48d7-9420-16d6f34fedf8\") " pod="openstack/cinder-df60-account-create-update-mlxn7" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.478749 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrqmv\" (UniqueName: \"kubernetes.io/projected/7ebabded-4179-4586-9be7-40d4635f7806-kube-api-access-mrqmv\") pod \"keystone-db-sync-5bhqp\" (UID: \"7ebabded-4179-4586-9be7-40d4635f7806\") " pod="openstack/keystone-db-sync-5bhqp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.478809 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebabded-4179-4586-9be7-40d4635f7806-combined-ca-bundle\") pod \"keystone-db-sync-5bhqp\" (UID: \"7ebabded-4179-4586-9be7-40d4635f7806\") " pod="openstack/keystone-db-sync-5bhqp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.478901 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebabded-4179-4586-9be7-40d4635f7806-config-data\") pod \"keystone-db-sync-5bhqp\" (UID: \"7ebabded-4179-4586-9be7-40d4635f7806\") " pod="openstack/keystone-db-sync-5bhqp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.478953 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01e65cec-01ec-4374-b9e7-c1b13f9b8d5e-operator-scripts\") pod \"neutron-db-create-dr57t\" (UID: \"01e65cec-01ec-4374-b9e7-c1b13f9b8d5e\") " pod="openstack/neutron-db-create-dr57t" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.478992 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm77g\" (UniqueName: \"kubernetes.io/projected/01e65cec-01ec-4374-b9e7-c1b13f9b8d5e-kube-api-access-lm77g\") pod \"neutron-db-create-dr57t\" (UID: \"01e65cec-01ec-4374-b9e7-c1b13f9b8d5e\") " pod="openstack/neutron-db-create-dr57t" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.486945 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-cb38-account-create-update-zdknp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.564883 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-df60-account-create-update-mlxn7" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.580934 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebabded-4179-4586-9be7-40d4635f7806-combined-ca-bundle\") pod \"keystone-db-sync-5bhqp\" (UID: \"7ebabded-4179-4586-9be7-40d4635f7806\") " pod="openstack/keystone-db-sync-5bhqp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.581047 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebabded-4179-4586-9be7-40d4635f7806-config-data\") pod \"keystone-db-sync-5bhqp\" (UID: \"7ebabded-4179-4586-9be7-40d4635f7806\") " pod="openstack/keystone-db-sync-5bhqp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.581112 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01e65cec-01ec-4374-b9e7-c1b13f9b8d5e-operator-scripts\") pod \"neutron-db-create-dr57t\" (UID: \"01e65cec-01ec-4374-b9e7-c1b13f9b8d5e\") " pod="openstack/neutron-db-create-dr57t" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.581158 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm77g\" (UniqueName: \"kubernetes.io/projected/01e65cec-01ec-4374-b9e7-c1b13f9b8d5e-kube-api-access-lm77g\") pod \"neutron-db-create-dr57t\" (UID: \"01e65cec-01ec-4374-b9e7-c1b13f9b8d5e\") " pod="openstack/neutron-db-create-dr57t" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.581219 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrqmv\" (UniqueName: \"kubernetes.io/projected/7ebabded-4179-4586-9be7-40d4635f7806-kube-api-access-mrqmv\") pod \"keystone-db-sync-5bhqp\" (UID: \"7ebabded-4179-4586-9be7-40d4635f7806\") " pod="openstack/keystone-db-sync-5bhqp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.585927 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01e65cec-01ec-4374-b9e7-c1b13f9b8d5e-operator-scripts\") pod \"neutron-db-create-dr57t\" (UID: \"01e65cec-01ec-4374-b9e7-c1b13f9b8d5e\") " pod="openstack/neutron-db-create-dr57t" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.592075 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebabded-4179-4586-9be7-40d4635f7806-combined-ca-bundle\") pod \"keystone-db-sync-5bhqp\" (UID: \"7ebabded-4179-4586-9be7-40d4635f7806\") " pod="openstack/keystone-db-sync-5bhqp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.593885 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebabded-4179-4586-9be7-40d4635f7806-config-data\") pod \"keystone-db-sync-5bhqp\" (UID: \"7ebabded-4179-4586-9be7-40d4635f7806\") " pod="openstack/keystone-db-sync-5bhqp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.595147 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d41e-account-create-update-jm4rk"] Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.600089 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d41e-account-create-update-jm4rk" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.630036 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d41e-account-create-update-jm4rk"] Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.631700 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.639536 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrqmv\" (UniqueName: \"kubernetes.io/projected/7ebabded-4179-4586-9be7-40d4635f7806-kube-api-access-mrqmv\") pod \"keystone-db-sync-5bhqp\" (UID: \"7ebabded-4179-4586-9be7-40d4635f7806\") " pod="openstack/keystone-db-sync-5bhqp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.658343 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm77g\" (UniqueName: \"kubernetes.io/projected/01e65cec-01ec-4374-b9e7-c1b13f9b8d5e-kube-api-access-lm77g\") pod \"neutron-db-create-dr57t\" (UID: \"01e65cec-01ec-4374-b9e7-c1b13f9b8d5e\") " pod="openstack/neutron-db-create-dr57t" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.689193 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw5fh\" (UniqueName: \"kubernetes.io/projected/c5df6e93-265c-4ff6-8bf5-974e1bc7c742-kube-api-access-rw5fh\") pod \"neutron-d41e-account-create-update-jm4rk\" (UID: \"c5df6e93-265c-4ff6-8bf5-974e1bc7c742\") " pod="openstack/neutron-d41e-account-create-update-jm4rk" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.689317 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5df6e93-265c-4ff6-8bf5-974e1bc7c742-operator-scripts\") pod \"neutron-d41e-account-create-update-jm4rk\" (UID: \"c5df6e93-265c-4ff6-8bf5-974e1bc7c742\") " pod="openstack/neutron-d41e-account-create-update-jm4rk" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.777183 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-5bhqp" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.784663 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dr57t" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.790923 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw5fh\" (UniqueName: \"kubernetes.io/projected/c5df6e93-265c-4ff6-8bf5-974e1bc7c742-kube-api-access-rw5fh\") pod \"neutron-d41e-account-create-update-jm4rk\" (UID: \"c5df6e93-265c-4ff6-8bf5-974e1bc7c742\") " pod="openstack/neutron-d41e-account-create-update-jm4rk" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.791032 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5df6e93-265c-4ff6-8bf5-974e1bc7c742-operator-scripts\") pod \"neutron-d41e-account-create-update-jm4rk\" (UID: \"c5df6e93-265c-4ff6-8bf5-974e1bc7c742\") " pod="openstack/neutron-d41e-account-create-update-jm4rk" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.791704 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5df6e93-265c-4ff6-8bf5-974e1bc7c742-operator-scripts\") pod \"neutron-d41e-account-create-update-jm4rk\" (UID: \"c5df6e93-265c-4ff6-8bf5-974e1bc7c742\") " pod="openstack/neutron-d41e-account-create-update-jm4rk" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.816350 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw5fh\" (UniqueName: \"kubernetes.io/projected/c5df6e93-265c-4ff6-8bf5-974e1bc7c742-kube-api-access-rw5fh\") pod \"neutron-d41e-account-create-update-jm4rk\" (UID: \"c5df6e93-265c-4ff6-8bf5-974e1bc7c742\") " pod="openstack/neutron-d41e-account-create-update-jm4rk" Nov 24 17:42:41 crc kubenswrapper[4808]: I1124 17:42:41.984774 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qg8hc"] Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.000400 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d41e-account-create-update-jm4rk" Nov 24 17:42:42 crc kubenswrapper[4808]: W1124 17:42:42.009890 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod180dfd5f_4d34_42fc_983d_90a54e945a15.slice/crio-1e997b70952633182c3064b0b68ae40ea2b9883fa72ebf30d257c00d7b8ef7fd WatchSource:0}: Error finding container 1e997b70952633182c3064b0b68ae40ea2b9883fa72ebf30d257c00d7b8ef7fd: Status 404 returned error can't find the container with id 1e997b70952633182c3064b0b68ae40ea2b9883fa72ebf30d257c00d7b8ef7fd Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.065197 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-b2sdj"] Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.192732 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-cb38-account-create-update-zdknp"] Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.280665 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-df60-account-create-update-mlxn7"] Nov 24 17:42:42 crc kubenswrapper[4808]: W1124 17:42:42.312268 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf113a2b1_1597_48d7_9420_16d6f34fedf8.slice/crio-7e9f96924df3d8d53b2044800f57d5facee0b43d05802cf1dc5090f299933131 WatchSource:0}: Error finding container 7e9f96924df3d8d53b2044800f57d5facee0b43d05802cf1dc5090f299933131: Status 404 returned error can't find the container with id 7e9f96924df3d8d53b2044800f57d5facee0b43d05802cf1dc5090f299933131 Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.533567 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-5bhqp"] Nov 24 17:42:42 crc kubenswrapper[4808]: W1124 17:42:42.534230 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ebabded_4179_4586_9be7_40d4635f7806.slice/crio-bf7461375e3bd04ae700689db4ad2874e88509a7a31e4b966a72f84ba763f02d WatchSource:0}: Error finding container bf7461375e3bd04ae700689db4ad2874e88509a7a31e4b966a72f84ba763f02d: Status 404 returned error can't find the container with id bf7461375e3bd04ae700689db4ad2874e88509a7a31e4b966a72f84ba763f02d Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.586925 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.621065 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-combined-ca-bundle\") pod \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.621153 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-db-sync-config-data\") pod \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.621175 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjx82\" (UniqueName: \"kubernetes.io/projected/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-kube-api-access-sjx82\") pod \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.621255 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-config-data\") pod \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\" (UID: \"813f14a2-3982-4e6d-aca7-109d2bf2d1cb\") " Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.641215 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-kube-api-access-sjx82" (OuterVolumeSpecName: "kube-api-access-sjx82") pod "813f14a2-3982-4e6d-aca7-109d2bf2d1cb" (UID: "813f14a2-3982-4e6d-aca7-109d2bf2d1cb"). InnerVolumeSpecName "kube-api-access-sjx82". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.649236 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "813f14a2-3982-4e6d-aca7-109d2bf2d1cb" (UID: "813f14a2-3982-4e6d-aca7-109d2bf2d1cb"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.663462 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-dr57t"] Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.701418 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-config-data" (OuterVolumeSpecName: "config-data") pod "813f14a2-3982-4e6d-aca7-109d2bf2d1cb" (UID: "813f14a2-3982-4e6d-aca7-109d2bf2d1cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.726274 4808 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.726317 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjx82\" (UniqueName: \"kubernetes.io/projected/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-kube-api-access-sjx82\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.726334 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.737063 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d41e-account-create-update-jm4rk"] Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.753121 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "813f14a2-3982-4e6d-aca7-109d2bf2d1cb" (UID: "813f14a2-3982-4e6d-aca7-109d2bf2d1cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.827604 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/813f14a2-3982-4e6d-aca7-109d2bf2d1cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.909945 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-d6rlt" Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.909940 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-d6rlt" event={"ID":"813f14a2-3982-4e6d-aca7-109d2bf2d1cb","Type":"ContainerDied","Data":"7132ebd5e3473eb27bc395888ab3e16cacbbe47f0da17deef297aebad7b51824"} Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.910156 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7132ebd5e3473eb27bc395888ab3e16cacbbe47f0da17deef297aebad7b51824" Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.911528 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dr57t" event={"ID":"01e65cec-01ec-4374-b9e7-c1b13f9b8d5e","Type":"ContainerStarted","Data":"52ed0d373b7440bd7b4bebe6c7bb72d50e7aa0b08704410979d013e188fa72d9"} Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.911565 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dr57t" event={"ID":"01e65cec-01ec-4374-b9e7-c1b13f9b8d5e","Type":"ContainerStarted","Data":"81ae4648116f8014ae1ddc7d684af2db210d66c038fc63d43469ee1e5327be05"} Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.912367 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d41e-account-create-update-jm4rk" event={"ID":"c5df6e93-265c-4ff6-8bf5-974e1bc7c742","Type":"ContainerStarted","Data":"57691ba68de79ab833c205d273963d6cc8739e38fad03c7214e806b1adba7066"} Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.913943 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-cb38-account-create-update-zdknp" event={"ID":"53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a","Type":"ContainerStarted","Data":"b44d74ea56f05888af192234ad258fc574a05e31916e6b4004f52c0fc779444b"} Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.914000 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-cb38-account-create-update-zdknp" event={"ID":"53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a","Type":"ContainerStarted","Data":"3175e2870a3625f4ad873aa1f085a4858bcc1691f51ace1c1a7d5d1fbe060c68"} Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.915003 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-b2sdj" event={"ID":"04e46a77-a53c-4edd-a1d8-c74a7b61c666","Type":"ContainerStarted","Data":"646d8e94f3759ceb3ec9c2a8ebd53163ad4acb5b28a241abafec5e92f2230717"} Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.915060 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-b2sdj" event={"ID":"04e46a77-a53c-4edd-a1d8-c74a7b61c666","Type":"ContainerStarted","Data":"053d6f474ed8daf999e49004373850ab8eef0c482e824cce21461df26a69f45e"} Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.916040 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qg8hc" event={"ID":"180dfd5f-4d34-42fc-983d-90a54e945a15","Type":"ContainerStarted","Data":"bda72b150f53176d32dabd034c30198f9b3df9da8544fb4ad46c3f00f23fc9b7"} Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.916070 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qg8hc" event={"ID":"180dfd5f-4d34-42fc-983d-90a54e945a15","Type":"ContainerStarted","Data":"1e997b70952633182c3064b0b68ae40ea2b9883fa72ebf30d257c00d7b8ef7fd"} Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.916973 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-df60-account-create-update-mlxn7" event={"ID":"f113a2b1-1597-48d7-9420-16d6f34fedf8","Type":"ContainerStarted","Data":"7c668f7132a032ed325d0ad3fd827286c17794f129ae67da35df367aaa4aaf55"} Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.917008 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-df60-account-create-update-mlxn7" event={"ID":"f113a2b1-1597-48d7-9420-16d6f34fedf8","Type":"ContainerStarted","Data":"7e9f96924df3d8d53b2044800f57d5facee0b43d05802cf1dc5090f299933131"} Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.918431 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-5bhqp" event={"ID":"7ebabded-4179-4586-9be7-40d4635f7806","Type":"ContainerStarted","Data":"bf7461375e3bd04ae700689db4ad2874e88509a7a31e4b966a72f84ba763f02d"} Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.935044 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-cb38-account-create-update-zdknp" podStartSLOduration=1.935028329 podStartE2EDuration="1.935028329s" podCreationTimestamp="2025-11-24 17:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:42:42.932450274 +0000 UTC m=+955.530118076" watchObservedRunningTime="2025-11-24 17:42:42.935028329 +0000 UTC m=+955.532696131" Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.961433 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-qg8hc" podStartSLOduration=2.961417387 podStartE2EDuration="2.961417387s" podCreationTimestamp="2025-11-24 17:42:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:42:42.957965247 +0000 UTC m=+955.555633059" watchObservedRunningTime="2025-11-24 17:42:42.961417387 +0000 UTC m=+955.559085189" Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.978829 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-df60-account-create-update-mlxn7" podStartSLOduration=1.978810233 podStartE2EDuration="1.978810233s" podCreationTimestamp="2025-11-24 17:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:42:42.975807866 +0000 UTC m=+955.573475668" watchObservedRunningTime="2025-11-24 17:42:42.978810233 +0000 UTC m=+955.576478035" Nov 24 17:42:42 crc kubenswrapper[4808]: I1124 17:42:42.998645 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-b2sdj" podStartSLOduration=1.9986218999999998 podStartE2EDuration="1.9986219s" podCreationTimestamp="2025-11-24 17:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:42:42.991272726 +0000 UTC m=+955.588940518" watchObservedRunningTime="2025-11-24 17:42:42.9986219 +0000 UTC m=+955.596289702" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.265075 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-2rpjx"] Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.265658 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" podUID="4cd2504d-284c-47ef-8579-78a798d4a5cb" containerName="dnsmasq-dns" containerID="cri-o://69fae8344fc7fd25f56c916847ee0d2dabdfe0165e8919776a3be809d2b527af" gracePeriod=10 Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.311283 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-xhk99"] Nov 24 17:42:43 crc kubenswrapper[4808]: E1124 17:42:43.311604 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="813f14a2-3982-4e6d-aca7-109d2bf2d1cb" containerName="glance-db-sync" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.311616 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="813f14a2-3982-4e6d-aca7-109d2bf2d1cb" containerName="glance-db-sync" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.311808 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="813f14a2-3982-4e6d-aca7-109d2bf2d1cb" containerName="glance-db-sync" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.315660 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.332575 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-xhk99"] Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.442176 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjh4b\" (UniqueName: \"kubernetes.io/projected/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-kube-api-access-pjh4b\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.442225 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.442251 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.442313 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.442386 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-config\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.442424 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.545466 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjh4b\" (UniqueName: \"kubernetes.io/projected/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-kube-api-access-pjh4b\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.545512 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.545532 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.545591 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.545631 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-config\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.545656 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.546387 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.546470 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.546540 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.546815 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.546939 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-config\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.583133 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjh4b\" (UniqueName: \"kubernetes.io/projected/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-kube-api-access-pjh4b\") pod \"dnsmasq-dns-74f6bcbc87-xhk99\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.777536 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.948279 4808 generic.go:334] "Generic (PLEG): container finished" podID="4cd2504d-284c-47ef-8579-78a798d4a5cb" containerID="69fae8344fc7fd25f56c916847ee0d2dabdfe0165e8919776a3be809d2b527af" exitCode=0 Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.948629 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" event={"ID":"4cd2504d-284c-47ef-8579-78a798d4a5cb","Type":"ContainerDied","Data":"69fae8344fc7fd25f56c916847ee0d2dabdfe0165e8919776a3be809d2b527af"} Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.956405 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d41e-account-create-update-jm4rk" event={"ID":"c5df6e93-265c-4ff6-8bf5-974e1bc7c742","Type":"ContainerStarted","Data":"2bd53b4eed9898815b0999e1fd92e3a9df0481a425074def41bd8f58099737d5"} Nov 24 17:42:43 crc kubenswrapper[4808]: I1124 17:42:43.995116 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-dr57t" podStartSLOduration=2.995099021 podStartE2EDuration="2.995099021s" podCreationTimestamp="2025-11-24 17:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:42:43.978869429 +0000 UTC m=+956.576537231" watchObservedRunningTime="2025-11-24 17:42:43.995099021 +0000 UTC m=+956.592766823" Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:44.021167 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-d41e-account-create-update-jm4rk" podStartSLOduration=3.021141678 podStartE2EDuration="3.021141678s" podCreationTimestamp="2025-11-24 17:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:42:44.010455317 +0000 UTC m=+956.608123119" watchObservedRunningTime="2025-11-24 17:42:44.021141678 +0000 UTC m=+956.618809490" Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:44.291524 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-xhk99"] Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:44.970769 4808 generic.go:334] "Generic (PLEG): container finished" podID="c5df6e93-265c-4ff6-8bf5-974e1bc7c742" containerID="2bd53b4eed9898815b0999e1fd92e3a9df0481a425074def41bd8f58099737d5" exitCode=0 Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:44.970980 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d41e-account-create-update-jm4rk" event={"ID":"c5df6e93-265c-4ff6-8bf5-974e1bc7c742","Type":"ContainerDied","Data":"2bd53b4eed9898815b0999e1fd92e3a9df0481a425074def41bd8f58099737d5"} Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:44.972943 4808 generic.go:334] "Generic (PLEG): container finished" podID="53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a" containerID="b44d74ea56f05888af192234ad258fc574a05e31916e6b4004f52c0fc779444b" exitCode=0 Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:44.972994 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-cb38-account-create-update-zdknp" event={"ID":"53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a","Type":"ContainerDied","Data":"b44d74ea56f05888af192234ad258fc574a05e31916e6b4004f52c0fc779444b"} Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:44.977557 4808 generic.go:334] "Generic (PLEG): container finished" podID="04e46a77-a53c-4edd-a1d8-c74a7b61c666" containerID="646d8e94f3759ceb3ec9c2a8ebd53163ad4acb5b28a241abafec5e92f2230717" exitCode=0 Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:44.977631 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-b2sdj" event={"ID":"04e46a77-a53c-4edd-a1d8-c74a7b61c666","Type":"ContainerDied","Data":"646d8e94f3759ceb3ec9c2a8ebd53163ad4acb5b28a241abafec5e92f2230717"} Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:44.991699 4808 generic.go:334] "Generic (PLEG): container finished" podID="45350630-8ea8-4ca6-83b1-f5a8cb26afb9" containerID="ba881a33e0592c1120cad1782d30e5127b95e61fef296581b977002ccd71649e" exitCode=0 Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:44.991875 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" event={"ID":"45350630-8ea8-4ca6-83b1-f5a8cb26afb9","Type":"ContainerDied","Data":"ba881a33e0592c1120cad1782d30e5127b95e61fef296581b977002ccd71649e"} Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:44.991903 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" event={"ID":"45350630-8ea8-4ca6-83b1-f5a8cb26afb9","Type":"ContainerStarted","Data":"cfb38c13e871e09c2000a534524cd6dc138bad9fd5d921ad0f6e5ef92269aed3"} Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:44.995833 4808 generic.go:334] "Generic (PLEG): container finished" podID="180dfd5f-4d34-42fc-983d-90a54e945a15" containerID="bda72b150f53176d32dabd034c30198f9b3df9da8544fb4ad46c3f00f23fc9b7" exitCode=0 Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:44.995922 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qg8hc" event={"ID":"180dfd5f-4d34-42fc-983d-90a54e945a15","Type":"ContainerDied","Data":"bda72b150f53176d32dabd034c30198f9b3df9da8544fb4ad46c3f00f23fc9b7"} Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:44.999301 4808 generic.go:334] "Generic (PLEG): container finished" podID="f113a2b1-1597-48d7-9420-16d6f34fedf8" containerID="7c668f7132a032ed325d0ad3fd827286c17794f129ae67da35df367aaa4aaf55" exitCode=0 Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:44.999358 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-df60-account-create-update-mlxn7" event={"ID":"f113a2b1-1597-48d7-9420-16d6f34fedf8","Type":"ContainerDied","Data":"7c668f7132a032ed325d0ad3fd827286c17794f129ae67da35df367aaa4aaf55"} Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.006173 4808 generic.go:334] "Generic (PLEG): container finished" podID="01e65cec-01ec-4374-b9e7-c1b13f9b8d5e" containerID="52ed0d373b7440bd7b4bebe6c7bb72d50e7aa0b08704410979d013e188fa72d9" exitCode=0 Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.006243 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dr57t" event={"ID":"01e65cec-01ec-4374-b9e7-c1b13f9b8d5e","Type":"ContainerDied","Data":"52ed0d373b7440bd7b4bebe6c7bb72d50e7aa0b08704410979d013e188fa72d9"} Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.264620 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.322706 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-dns-swift-storage-0\") pod \"4cd2504d-284c-47ef-8579-78a798d4a5cb\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.323062 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-dns-svc\") pod \"4cd2504d-284c-47ef-8579-78a798d4a5cb\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.323151 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-config\") pod \"4cd2504d-284c-47ef-8579-78a798d4a5cb\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.323194 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-ovsdbserver-sb\") pod \"4cd2504d-284c-47ef-8579-78a798d4a5cb\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.323257 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-ovsdbserver-nb\") pod \"4cd2504d-284c-47ef-8579-78a798d4a5cb\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.323320 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b44qk\" (UniqueName: \"kubernetes.io/projected/4cd2504d-284c-47ef-8579-78a798d4a5cb-kube-api-access-b44qk\") pod \"4cd2504d-284c-47ef-8579-78a798d4a5cb\" (UID: \"4cd2504d-284c-47ef-8579-78a798d4a5cb\") " Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.330575 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cd2504d-284c-47ef-8579-78a798d4a5cb-kube-api-access-b44qk" (OuterVolumeSpecName: "kube-api-access-b44qk") pod "4cd2504d-284c-47ef-8579-78a798d4a5cb" (UID: "4cd2504d-284c-47ef-8579-78a798d4a5cb"). InnerVolumeSpecName "kube-api-access-b44qk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.387922 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4cd2504d-284c-47ef-8579-78a798d4a5cb" (UID: "4cd2504d-284c-47ef-8579-78a798d4a5cb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.390277 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4cd2504d-284c-47ef-8579-78a798d4a5cb" (UID: "4cd2504d-284c-47ef-8579-78a798d4a5cb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.397298 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-config" (OuterVolumeSpecName: "config") pod "4cd2504d-284c-47ef-8579-78a798d4a5cb" (UID: "4cd2504d-284c-47ef-8579-78a798d4a5cb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.398335 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4cd2504d-284c-47ef-8579-78a798d4a5cb" (UID: "4cd2504d-284c-47ef-8579-78a798d4a5cb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.417435 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4cd2504d-284c-47ef-8579-78a798d4a5cb" (UID: "4cd2504d-284c-47ef-8579-78a798d4a5cb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.425325 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.425357 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b44qk\" (UniqueName: \"kubernetes.io/projected/4cd2504d-284c-47ef-8579-78a798d4a5cb-kube-api-access-b44qk\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.425370 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.425378 4808 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.425387 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:45 crc kubenswrapper[4808]: I1124 17:42:45.425395 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cd2504d-284c-47ef-8579-78a798d4a5cb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:46 crc kubenswrapper[4808]: I1124 17:42:46.016861 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" event={"ID":"4cd2504d-284c-47ef-8579-78a798d4a5cb","Type":"ContainerDied","Data":"7f2999bf27f145294eda83e7d6802fe0cdae9e963e01ff95979c0623f8200927"} Nov 24 17:42:46 crc kubenswrapper[4808]: I1124 17:42:46.017309 4808 scope.go:117] "RemoveContainer" containerID="69fae8344fc7fd25f56c916847ee0d2dabdfe0165e8919776a3be809d2b527af" Nov 24 17:42:46 crc kubenswrapper[4808]: I1124 17:42:46.016891 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-2rpjx" Nov 24 17:42:46 crc kubenswrapper[4808]: I1124 17:42:46.020683 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" event={"ID":"45350630-8ea8-4ca6-83b1-f5a8cb26afb9","Type":"ContainerStarted","Data":"726ba42ada2711d0602c876e9abe812da88a44225ec61712c953d52f5c328479"} Nov 24 17:42:46 crc kubenswrapper[4808]: I1124 17:42:46.044866 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" podStartSLOduration=3.044851205 podStartE2EDuration="3.044851205s" podCreationTimestamp="2025-11-24 17:42:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:42:46.041720653 +0000 UTC m=+958.639388455" watchObservedRunningTime="2025-11-24 17:42:46.044851205 +0000 UTC m=+958.642519007" Nov 24 17:42:46 crc kubenswrapper[4808]: I1124 17:42:46.065876 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-2rpjx"] Nov 24 17:42:46 crc kubenswrapper[4808]: I1124 17:42:46.086289 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-2rpjx"] Nov 24 17:42:46 crc kubenswrapper[4808]: I1124 17:42:46.357682 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cd2504d-284c-47ef-8579-78a798d4a5cb" path="/var/lib/kubelet/pods/4cd2504d-284c-47ef-8579-78a798d4a5cb/volumes" Nov 24 17:42:47 crc kubenswrapper[4808]: I1124 17:42:47.028192 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.287985 4808 scope.go:117] "RemoveContainer" containerID="6767976c7a11d87f388fa4e6cce062addc10ada3f5582a76073742371b53a0b3" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.510945 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qg8hc" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.588186 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/180dfd5f-4d34-42fc-983d-90a54e945a15-operator-scripts\") pod \"180dfd5f-4d34-42fc-983d-90a54e945a15\" (UID: \"180dfd5f-4d34-42fc-983d-90a54e945a15\") " Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.588352 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjzq4\" (UniqueName: \"kubernetes.io/projected/180dfd5f-4d34-42fc-983d-90a54e945a15-kube-api-access-bjzq4\") pod \"180dfd5f-4d34-42fc-983d-90a54e945a15\" (UID: \"180dfd5f-4d34-42fc-983d-90a54e945a15\") " Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.588922 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/180dfd5f-4d34-42fc-983d-90a54e945a15-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "180dfd5f-4d34-42fc-983d-90a54e945a15" (UID: "180dfd5f-4d34-42fc-983d-90a54e945a15"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.593371 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/180dfd5f-4d34-42fc-983d-90a54e945a15-kube-api-access-bjzq4" (OuterVolumeSpecName: "kube-api-access-bjzq4") pod "180dfd5f-4d34-42fc-983d-90a54e945a15" (UID: "180dfd5f-4d34-42fc-983d-90a54e945a15"). InnerVolumeSpecName "kube-api-access-bjzq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.607978 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d41e-account-create-update-jm4rk" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.616784 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dr57t" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.656448 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b2sdj" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.668295 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-df60-account-create-update-mlxn7" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.674671 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-cb38-account-create-update-zdknp" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.689470 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01e65cec-01ec-4374-b9e7-c1b13f9b8d5e-operator-scripts\") pod \"01e65cec-01ec-4374-b9e7-c1b13f9b8d5e\" (UID: \"01e65cec-01ec-4374-b9e7-c1b13f9b8d5e\") " Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.689568 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5df6e93-265c-4ff6-8bf5-974e1bc7c742-operator-scripts\") pod \"c5df6e93-265c-4ff6-8bf5-974e1bc7c742\" (UID: \"c5df6e93-265c-4ff6-8bf5-974e1bc7c742\") " Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.689662 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm77g\" (UniqueName: \"kubernetes.io/projected/01e65cec-01ec-4374-b9e7-c1b13f9b8d5e-kube-api-access-lm77g\") pod \"01e65cec-01ec-4374-b9e7-c1b13f9b8d5e\" (UID: \"01e65cec-01ec-4374-b9e7-c1b13f9b8d5e\") " Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.689786 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw5fh\" (UniqueName: \"kubernetes.io/projected/c5df6e93-265c-4ff6-8bf5-974e1bc7c742-kube-api-access-rw5fh\") pod \"c5df6e93-265c-4ff6-8bf5-974e1bc7c742\" (UID: \"c5df6e93-265c-4ff6-8bf5-974e1bc7c742\") " Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.690238 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjzq4\" (UniqueName: \"kubernetes.io/projected/180dfd5f-4d34-42fc-983d-90a54e945a15-kube-api-access-bjzq4\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.690261 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/180dfd5f-4d34-42fc-983d-90a54e945a15-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.690658 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01e65cec-01ec-4374-b9e7-c1b13f9b8d5e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "01e65cec-01ec-4374-b9e7-c1b13f9b8d5e" (UID: "01e65cec-01ec-4374-b9e7-c1b13f9b8d5e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.691027 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5df6e93-265c-4ff6-8bf5-974e1bc7c742-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c5df6e93-265c-4ff6-8bf5-974e1bc7c742" (UID: "c5df6e93-265c-4ff6-8bf5-974e1bc7c742"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.694759 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01e65cec-01ec-4374-b9e7-c1b13f9b8d5e-kube-api-access-lm77g" (OuterVolumeSpecName: "kube-api-access-lm77g") pod "01e65cec-01ec-4374-b9e7-c1b13f9b8d5e" (UID: "01e65cec-01ec-4374-b9e7-c1b13f9b8d5e"). InnerVolumeSpecName "kube-api-access-lm77g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.697202 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5df6e93-265c-4ff6-8bf5-974e1bc7c742-kube-api-access-rw5fh" (OuterVolumeSpecName: "kube-api-access-rw5fh") pod "c5df6e93-265c-4ff6-8bf5-974e1bc7c742" (UID: "c5df6e93-265c-4ff6-8bf5-974e1bc7c742"). InnerVolumeSpecName "kube-api-access-rw5fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.791370 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a-operator-scripts\") pod \"53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a\" (UID: \"53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a\") " Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.791719 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9rfz\" (UniqueName: \"kubernetes.io/projected/53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a-kube-api-access-r9rfz\") pod \"53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a\" (UID: \"53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a\") " Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.791767 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vv7kd\" (UniqueName: \"kubernetes.io/projected/f113a2b1-1597-48d7-9420-16d6f34fedf8-kube-api-access-vv7kd\") pod \"f113a2b1-1597-48d7-9420-16d6f34fedf8\" (UID: \"f113a2b1-1597-48d7-9420-16d6f34fedf8\") " Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.791797 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f113a2b1-1597-48d7-9420-16d6f34fedf8-operator-scripts\") pod \"f113a2b1-1597-48d7-9420-16d6f34fedf8\" (UID: \"f113a2b1-1597-48d7-9420-16d6f34fedf8\") " Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.791843 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04e46a77-a53c-4edd-a1d8-c74a7b61c666-operator-scripts\") pod \"04e46a77-a53c-4edd-a1d8-c74a7b61c666\" (UID: \"04e46a77-a53c-4edd-a1d8-c74a7b61c666\") " Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.791938 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9jv9\" (UniqueName: \"kubernetes.io/projected/04e46a77-a53c-4edd-a1d8-c74a7b61c666-kube-api-access-j9jv9\") pod \"04e46a77-a53c-4edd-a1d8-c74a7b61c666\" (UID: \"04e46a77-a53c-4edd-a1d8-c74a7b61c666\") " Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.792171 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a" (UID: "53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.792508 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.792919 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm77g\" (UniqueName: \"kubernetes.io/projected/01e65cec-01ec-4374-b9e7-c1b13f9b8d5e-kube-api-access-lm77g\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.792953 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw5fh\" (UniqueName: \"kubernetes.io/projected/c5df6e93-265c-4ff6-8bf5-974e1bc7c742-kube-api-access-rw5fh\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.792964 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01e65cec-01ec-4374-b9e7-c1b13f9b8d5e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.792973 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5df6e93-265c-4ff6-8bf5-974e1bc7c742-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.792531 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f113a2b1-1597-48d7-9420-16d6f34fedf8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f113a2b1-1597-48d7-9420-16d6f34fedf8" (UID: "f113a2b1-1597-48d7-9420-16d6f34fedf8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.792711 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04e46a77-a53c-4edd-a1d8-c74a7b61c666-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "04e46a77-a53c-4edd-a1d8-c74a7b61c666" (UID: "04e46a77-a53c-4edd-a1d8-c74a7b61c666"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.795893 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04e46a77-a53c-4edd-a1d8-c74a7b61c666-kube-api-access-j9jv9" (OuterVolumeSpecName: "kube-api-access-j9jv9") pod "04e46a77-a53c-4edd-a1d8-c74a7b61c666" (UID: "04e46a77-a53c-4edd-a1d8-c74a7b61c666"). InnerVolumeSpecName "kube-api-access-j9jv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.796388 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a-kube-api-access-r9rfz" (OuterVolumeSpecName: "kube-api-access-r9rfz") pod "53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a" (UID: "53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a"). InnerVolumeSpecName "kube-api-access-r9rfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.799348 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f113a2b1-1597-48d7-9420-16d6f34fedf8-kube-api-access-vv7kd" (OuterVolumeSpecName: "kube-api-access-vv7kd") pod "f113a2b1-1597-48d7-9420-16d6f34fedf8" (UID: "f113a2b1-1597-48d7-9420-16d6f34fedf8"). InnerVolumeSpecName "kube-api-access-vv7kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.894556 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9jv9\" (UniqueName: \"kubernetes.io/projected/04e46a77-a53c-4edd-a1d8-c74a7b61c666-kube-api-access-j9jv9\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.894579 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9rfz\" (UniqueName: \"kubernetes.io/projected/53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a-kube-api-access-r9rfz\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.894590 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vv7kd\" (UniqueName: \"kubernetes.io/projected/f113a2b1-1597-48d7-9420-16d6f34fedf8-kube-api-access-vv7kd\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.894599 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f113a2b1-1597-48d7-9420-16d6f34fedf8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:48 crc kubenswrapper[4808]: I1124 17:42:48.894608 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04e46a77-a53c-4edd-a1d8-c74a7b61c666-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.048684 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-5bhqp" event={"ID":"7ebabded-4179-4586-9be7-40d4635f7806","Type":"ContainerStarted","Data":"afa0a9324c77e5234ce460beb02ca8d9614289ad4644a4244b1a739e7c1072f4"} Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.050805 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dr57t" Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.050986 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dr57t" event={"ID":"01e65cec-01ec-4374-b9e7-c1b13f9b8d5e","Type":"ContainerDied","Data":"81ae4648116f8014ae1ddc7d684af2db210d66c038fc63d43469ee1e5327be05"} Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.051152 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81ae4648116f8014ae1ddc7d684af2db210d66c038fc63d43469ee1e5327be05" Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.053177 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d41e-account-create-update-jm4rk" event={"ID":"c5df6e93-265c-4ff6-8bf5-974e1bc7c742","Type":"ContainerDied","Data":"57691ba68de79ab833c205d273963d6cc8739e38fad03c7214e806b1adba7066"} Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.053223 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57691ba68de79ab833c205d273963d6cc8739e38fad03c7214e806b1adba7066" Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.053313 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d41e-account-create-update-jm4rk" Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.059372 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-cb38-account-create-update-zdknp" event={"ID":"53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a","Type":"ContainerDied","Data":"3175e2870a3625f4ad873aa1f085a4858bcc1691f51ace1c1a7d5d1fbe060c68"} Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.059412 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-cb38-account-create-update-zdknp" Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.059422 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3175e2870a3625f4ad873aa1f085a4858bcc1691f51ace1c1a7d5d1fbe060c68" Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.061397 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-b2sdj" event={"ID":"04e46a77-a53c-4edd-a1d8-c74a7b61c666","Type":"ContainerDied","Data":"053d6f474ed8daf999e49004373850ab8eef0c482e824cce21461df26a69f45e"} Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.061505 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="053d6f474ed8daf999e49004373850ab8eef0c482e824cce21461df26a69f45e" Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.061668 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b2sdj" Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.067550 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qg8hc" event={"ID":"180dfd5f-4d34-42fc-983d-90a54e945a15","Type":"ContainerDied","Data":"1e997b70952633182c3064b0b68ae40ea2b9883fa72ebf30d257c00d7b8ef7fd"} Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.067597 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e997b70952633182c3064b0b68ae40ea2b9883fa72ebf30d257c00d7b8ef7fd" Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.067594 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qg8hc" Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.069150 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-5bhqp" podStartSLOduration=2.2160464539999998 podStartE2EDuration="8.069129861s" podCreationTimestamp="2025-11-24 17:42:41 +0000 UTC" firstStartedPulling="2025-11-24 17:42:42.53669095 +0000 UTC m=+955.134358752" lastFinishedPulling="2025-11-24 17:42:48.389774357 +0000 UTC m=+960.987442159" observedRunningTime="2025-11-24 17:42:49.066177375 +0000 UTC m=+961.663845177" watchObservedRunningTime="2025-11-24 17:42:49.069129861 +0000 UTC m=+961.666797663" Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.069456 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-df60-account-create-update-mlxn7" event={"ID":"f113a2b1-1597-48d7-9420-16d6f34fedf8","Type":"ContainerDied","Data":"7e9f96924df3d8d53b2044800f57d5facee0b43d05802cf1dc5090f299933131"} Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.069497 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e9f96924df3d8d53b2044800f57d5facee0b43d05802cf1dc5090f299933131" Nov 24 17:42:49 crc kubenswrapper[4808]: I1124 17:42:49.069476 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-df60-account-create-update-mlxn7" Nov 24 17:42:52 crc kubenswrapper[4808]: I1124 17:42:52.115088 4808 generic.go:334] "Generic (PLEG): container finished" podID="7ebabded-4179-4586-9be7-40d4635f7806" containerID="afa0a9324c77e5234ce460beb02ca8d9614289ad4644a4244b1a739e7c1072f4" exitCode=0 Nov 24 17:42:52 crc kubenswrapper[4808]: I1124 17:42:52.115204 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-5bhqp" event={"ID":"7ebabded-4179-4586-9be7-40d4635f7806","Type":"ContainerDied","Data":"afa0a9324c77e5234ce460beb02ca8d9614289ad4644a4244b1a739e7c1072f4"} Nov 24 17:42:53 crc kubenswrapper[4808]: I1124 17:42:53.428546 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-5bhqp" Nov 24 17:42:53 crc kubenswrapper[4808]: I1124 17:42:53.562063 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrqmv\" (UniqueName: \"kubernetes.io/projected/7ebabded-4179-4586-9be7-40d4635f7806-kube-api-access-mrqmv\") pod \"7ebabded-4179-4586-9be7-40d4635f7806\" (UID: \"7ebabded-4179-4586-9be7-40d4635f7806\") " Nov 24 17:42:53 crc kubenswrapper[4808]: I1124 17:42:53.562146 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebabded-4179-4586-9be7-40d4635f7806-combined-ca-bundle\") pod \"7ebabded-4179-4586-9be7-40d4635f7806\" (UID: \"7ebabded-4179-4586-9be7-40d4635f7806\") " Nov 24 17:42:53 crc kubenswrapper[4808]: I1124 17:42:53.562209 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebabded-4179-4586-9be7-40d4635f7806-config-data\") pod \"7ebabded-4179-4586-9be7-40d4635f7806\" (UID: \"7ebabded-4179-4586-9be7-40d4635f7806\") " Nov 24 17:42:53 crc kubenswrapper[4808]: I1124 17:42:53.567156 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ebabded-4179-4586-9be7-40d4635f7806-kube-api-access-mrqmv" (OuterVolumeSpecName: "kube-api-access-mrqmv") pod "7ebabded-4179-4586-9be7-40d4635f7806" (UID: "7ebabded-4179-4586-9be7-40d4635f7806"). InnerVolumeSpecName "kube-api-access-mrqmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:53 crc kubenswrapper[4808]: I1124 17:42:53.587681 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ebabded-4179-4586-9be7-40d4635f7806-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ebabded-4179-4586-9be7-40d4635f7806" (UID: "7ebabded-4179-4586-9be7-40d4635f7806"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:42:53 crc kubenswrapper[4808]: I1124 17:42:53.610637 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ebabded-4179-4586-9be7-40d4635f7806-config-data" (OuterVolumeSpecName: "config-data") pod "7ebabded-4179-4586-9be7-40d4635f7806" (UID: "7ebabded-4179-4586-9be7-40d4635f7806"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:42:53 crc kubenswrapper[4808]: I1124 17:42:53.664669 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrqmv\" (UniqueName: \"kubernetes.io/projected/7ebabded-4179-4586-9be7-40d4635f7806-kube-api-access-mrqmv\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:53 crc kubenswrapper[4808]: I1124 17:42:53.664709 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebabded-4179-4586-9be7-40d4635f7806-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:53 crc kubenswrapper[4808]: I1124 17:42:53.664719 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebabded-4179-4586-9be7-40d4635f7806-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:53 crc kubenswrapper[4808]: I1124 17:42:53.780273 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:42:53 crc kubenswrapper[4808]: I1124 17:42:53.855185 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-f2jqs"] Nov 24 17:42:53 crc kubenswrapper[4808]: I1124 17:42:53.855440 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-f2jqs" podUID="328bde2b-812b-4cf8-b8a4-438ccbf6d108" containerName="dnsmasq-dns" containerID="cri-o://4b54deb9d465b1902512df3a773a8a3e978f9e9fbf6ab948b077fb78427ec67a" gracePeriod=10 Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.131395 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-5bhqp" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.131384 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-5bhqp" event={"ID":"7ebabded-4179-4586-9be7-40d4635f7806","Type":"ContainerDied","Data":"bf7461375e3bd04ae700689db4ad2874e88509a7a31e4b966a72f84ba763f02d"} Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.131960 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf7461375e3bd04ae700689db4ad2874e88509a7a31e4b966a72f84ba763f02d" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.133812 4808 generic.go:334] "Generic (PLEG): container finished" podID="328bde2b-812b-4cf8-b8a4-438ccbf6d108" containerID="4b54deb9d465b1902512df3a773a8a3e978f9e9fbf6ab948b077fb78427ec67a" exitCode=0 Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.133856 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-f2jqs" event={"ID":"328bde2b-812b-4cf8-b8a4-438ccbf6d108","Type":"ContainerDied","Data":"4b54deb9d465b1902512df3a773a8a3e978f9e9fbf6ab948b077fb78427ec67a"} Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.298312 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-5cl9j"] Nov 24 17:42:54 crc kubenswrapper[4808]: E1124 17:42:54.298724 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cd2504d-284c-47ef-8579-78a798d4a5cb" containerName="init" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.298748 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cd2504d-284c-47ef-8579-78a798d4a5cb" containerName="init" Nov 24 17:42:54 crc kubenswrapper[4808]: E1124 17:42:54.298777 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a" containerName="mariadb-account-create-update" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.298786 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a" containerName="mariadb-account-create-update" Nov 24 17:42:54 crc kubenswrapper[4808]: E1124 17:42:54.298801 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ebabded-4179-4586-9be7-40d4635f7806" containerName="keystone-db-sync" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.298809 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ebabded-4179-4586-9be7-40d4635f7806" containerName="keystone-db-sync" Nov 24 17:42:54 crc kubenswrapper[4808]: E1124 17:42:54.298819 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f113a2b1-1597-48d7-9420-16d6f34fedf8" containerName="mariadb-account-create-update" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.298826 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f113a2b1-1597-48d7-9420-16d6f34fedf8" containerName="mariadb-account-create-update" Nov 24 17:42:54 crc kubenswrapper[4808]: E1124 17:42:54.298834 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="180dfd5f-4d34-42fc-983d-90a54e945a15" containerName="mariadb-database-create" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.298841 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="180dfd5f-4d34-42fc-983d-90a54e945a15" containerName="mariadb-database-create" Nov 24 17:42:54 crc kubenswrapper[4808]: E1124 17:42:54.298851 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cd2504d-284c-47ef-8579-78a798d4a5cb" containerName="dnsmasq-dns" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.298860 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cd2504d-284c-47ef-8579-78a798d4a5cb" containerName="dnsmasq-dns" Nov 24 17:42:54 crc kubenswrapper[4808]: E1124 17:42:54.298875 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e65cec-01ec-4374-b9e7-c1b13f9b8d5e" containerName="mariadb-database-create" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.298883 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e65cec-01ec-4374-b9e7-c1b13f9b8d5e" containerName="mariadb-database-create" Nov 24 17:42:54 crc kubenswrapper[4808]: E1124 17:42:54.298894 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5df6e93-265c-4ff6-8bf5-974e1bc7c742" containerName="mariadb-account-create-update" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.298904 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5df6e93-265c-4ff6-8bf5-974e1bc7c742" containerName="mariadb-account-create-update" Nov 24 17:42:54 crc kubenswrapper[4808]: E1124 17:42:54.298917 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04e46a77-a53c-4edd-a1d8-c74a7b61c666" containerName="mariadb-database-create" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.298925 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="04e46a77-a53c-4edd-a1d8-c74a7b61c666" containerName="mariadb-database-create" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.299167 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="04e46a77-a53c-4edd-a1d8-c74a7b61c666" containerName="mariadb-database-create" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.299180 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="180dfd5f-4d34-42fc-983d-90a54e945a15" containerName="mariadb-database-create" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.299194 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cd2504d-284c-47ef-8579-78a798d4a5cb" containerName="dnsmasq-dns" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.299203 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="01e65cec-01ec-4374-b9e7-c1b13f9b8d5e" containerName="mariadb-database-create" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.299216 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5df6e93-265c-4ff6-8bf5-974e1bc7c742" containerName="mariadb-account-create-update" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.299227 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f113a2b1-1597-48d7-9420-16d6f34fedf8" containerName="mariadb-account-create-update" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.299235 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a" containerName="mariadb-account-create-update" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.299246 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ebabded-4179-4586-9be7-40d4635f7806" containerName="keystone-db-sync" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.300047 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.310926 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-5cl9j"] Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.340500 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-bfff8"] Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.341575 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.348411 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.348591 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.348620 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.348764 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-622rb" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.349664 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.370248 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bfff8"] Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.379867 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-config\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.379934 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-dns-svc\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.379960 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.380108 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.380220 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.380374 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw82n\" (UniqueName: \"kubernetes.io/projected/19a0a600-f3b4-4c6f-90bb-2889ba076149-kube-api-access-zw82n\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.481937 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw82n\" (UniqueName: \"kubernetes.io/projected/19a0a600-f3b4-4c6f-90bb-2889ba076149-kube-api-access-zw82n\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.482042 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-scripts\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.482079 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-credential-keys\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.482099 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-config\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.482159 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-combined-ca-bundle\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.482186 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-config-data\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.482223 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-dns-svc\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.482245 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-fernet-keys\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.482262 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.482284 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clm7p\" (UniqueName: \"kubernetes.io/projected/7dff84dc-26ca-4e8f-8386-37abe2041d7f-kube-api-access-clm7p\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.482305 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.482366 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.484432 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-dns-svc\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.491958 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-config\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.494726 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.495781 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.496577 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.507804 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw82n\" (UniqueName: \"kubernetes.io/projected/19a0a600-f3b4-4c6f-90bb-2889ba076149-kube-api-access-zw82n\") pod \"dnsmasq-dns-847c4cc679-5cl9j\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.538646 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-f886bf6d5-mrvf5"] Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.540457 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.545055 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-p77zx"] Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.547675 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.548364 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.548476 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-qdndj" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.548591 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.548700 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.551487 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.551664 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.551767 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-tjvqn" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.566312 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f886bf6d5-mrvf5"] Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.586455 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-scripts\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.586522 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-credential-keys\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.586578 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-combined-ca-bundle\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.586605 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-config-data\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.586646 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-fernet-keys\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.586670 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clm7p\" (UniqueName: \"kubernetes.io/projected/7dff84dc-26ca-4e8f-8386-37abe2041d7f-kube-api-access-clm7p\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.597972 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-config-data\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.598334 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-scripts\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.607458 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-credential-keys\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.611766 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-combined-ca-bundle\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.628239 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-fernet-keys\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.628730 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.633770 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-p77zx"] Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.662478 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clm7p\" (UniqueName: \"kubernetes.io/projected/7dff84dc-26ca-4e8f-8386-37abe2041d7f-kube-api-access-clm7p\") pod \"keystone-bootstrap-bfff8\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.691008 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e378c94e-0dd2-482a-bb40-6a965270a5a3-scripts\") pod \"horizon-f886bf6d5-mrvf5\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.691083 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-config-data\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.691115 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e378c94e-0dd2-482a-bb40-6a965270a5a3-config-data\") pod \"horizon-f886bf6d5-mrvf5\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.691143 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-527cr\" (UniqueName: \"kubernetes.io/projected/e378c94e-0dd2-482a-bb40-6a965270a5a3-kube-api-access-527cr\") pod \"horizon-f886bf6d5-mrvf5\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.691193 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-db-sync-config-data\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.691221 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b80d86dd-fbd8-444a-8159-d5c870eb01be-etc-machine-id\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.691243 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e378c94e-0dd2-482a-bb40-6a965270a5a3-logs\") pod \"horizon-f886bf6d5-mrvf5\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.691262 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-scripts\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.691281 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e378c94e-0dd2-482a-bb40-6a965270a5a3-horizon-secret-key\") pod \"horizon-f886bf6d5-mrvf5\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.691315 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l74k\" (UniqueName: \"kubernetes.io/projected/b80d86dd-fbd8-444a-8159-d5c870eb01be-kube-api-access-4l74k\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.691377 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-combined-ca-bundle\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.726279 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-kp6v5"] Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.726489 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.736397 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kp6v5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.767041 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-nspql" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.767287 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.767834 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.804299 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b2112bdd-0811-4bd0-927d-277b44b5cfc0-config\") pod \"neutron-db-sync-kp6v5\" (UID: \"b2112bdd-0811-4bd0-927d-277b44b5cfc0\") " pod="openstack/neutron-db-sync-kp6v5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.804375 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e378c94e-0dd2-482a-bb40-6a965270a5a3-scripts\") pod \"horizon-f886bf6d5-mrvf5\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.804416 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-config-data\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.804452 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e378c94e-0dd2-482a-bb40-6a965270a5a3-config-data\") pod \"horizon-f886bf6d5-mrvf5\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.804487 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-527cr\" (UniqueName: \"kubernetes.io/projected/e378c94e-0dd2-482a-bb40-6a965270a5a3-kube-api-access-527cr\") pod \"horizon-f886bf6d5-mrvf5\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.804538 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c654c\" (UniqueName: \"kubernetes.io/projected/b2112bdd-0811-4bd0-927d-277b44b5cfc0-kube-api-access-c654c\") pod \"neutron-db-sync-kp6v5\" (UID: \"b2112bdd-0811-4bd0-927d-277b44b5cfc0\") " pod="openstack/neutron-db-sync-kp6v5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.804590 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-db-sync-config-data\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.804621 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b80d86dd-fbd8-444a-8159-d5c870eb01be-etc-machine-id\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.804644 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e378c94e-0dd2-482a-bb40-6a965270a5a3-logs\") pod \"horizon-f886bf6d5-mrvf5\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.804662 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-scripts\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.804679 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e378c94e-0dd2-482a-bb40-6a965270a5a3-horizon-secret-key\") pod \"horizon-f886bf6d5-mrvf5\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.804699 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2112bdd-0811-4bd0-927d-277b44b5cfc0-combined-ca-bundle\") pod \"neutron-db-sync-kp6v5\" (UID: \"b2112bdd-0811-4bd0-927d-277b44b5cfc0\") " pod="openstack/neutron-db-sync-kp6v5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.804731 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l74k\" (UniqueName: \"kubernetes.io/projected/b80d86dd-fbd8-444a-8159-d5c870eb01be-kube-api-access-4l74k\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.804803 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-combined-ca-bundle\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.808363 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e378c94e-0dd2-482a-bb40-6a965270a5a3-scripts\") pod \"horizon-f886bf6d5-mrvf5\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.817171 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e378c94e-0dd2-482a-bb40-6a965270a5a3-config-data\") pod \"horizon-f886bf6d5-mrvf5\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.821182 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b80d86dd-fbd8-444a-8159-d5c870eb01be-etc-machine-id\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.821752 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e378c94e-0dd2-482a-bb40-6a965270a5a3-logs\") pod \"horizon-f886bf6d5-mrvf5\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.826553 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-db-sync-config-data\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.826676 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e378c94e-0dd2-482a-bb40-6a965270a5a3-horizon-secret-key\") pod \"horizon-f886bf6d5-mrvf5\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.829838 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-config-data\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.826988 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-scripts\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.834721 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-combined-ca-bundle\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.848815 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l74k\" (UniqueName: \"kubernetes.io/projected/b80d86dd-fbd8-444a-8159-d5c870eb01be-kube-api-access-4l74k\") pod \"cinder-db-sync-p77zx\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.852626 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.914549 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c654c\" (UniqueName: \"kubernetes.io/projected/b2112bdd-0811-4bd0-927d-277b44b5cfc0-kube-api-access-c654c\") pod \"neutron-db-sync-kp6v5\" (UID: \"b2112bdd-0811-4bd0-927d-277b44b5cfc0\") " pod="openstack/neutron-db-sync-kp6v5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.914624 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2112bdd-0811-4bd0-927d-277b44b5cfc0-combined-ca-bundle\") pod \"neutron-db-sync-kp6v5\" (UID: \"b2112bdd-0811-4bd0-927d-277b44b5cfc0\") " pod="openstack/neutron-db-sync-kp6v5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.914703 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b2112bdd-0811-4bd0-927d-277b44b5cfc0-config\") pod \"neutron-db-sync-kp6v5\" (UID: \"b2112bdd-0811-4bd0-927d-277b44b5cfc0\") " pod="openstack/neutron-db-sync-kp6v5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.909723 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-527cr\" (UniqueName: \"kubernetes.io/projected/e378c94e-0dd2-482a-bb40-6a965270a5a3-kube-api-access-527cr\") pod \"horizon-f886bf6d5-mrvf5\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.931774 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b2112bdd-0811-4bd0-927d-277b44b5cfc0-config\") pod \"neutron-db-sync-kp6v5\" (UID: \"b2112bdd-0811-4bd0-927d-277b44b5cfc0\") " pod="openstack/neutron-db-sync-kp6v5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.933890 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2112bdd-0811-4bd0-927d-277b44b5cfc0-combined-ca-bundle\") pod \"neutron-db-sync-kp6v5\" (UID: \"b2112bdd-0811-4bd0-927d-277b44b5cfc0\") " pod="openstack/neutron-db-sync-kp6v5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.934474 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.944868 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.960541 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.981166 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kp6v5"] Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.984982 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c654c\" (UniqueName: \"kubernetes.io/projected/b2112bdd-0811-4bd0-927d-277b44b5cfc0-kube-api-access-c654c\") pod \"neutron-db-sync-kp6v5\" (UID: \"b2112bdd-0811-4bd0-927d-277b44b5cfc0\") " pod="openstack/neutron-db-sync-kp6v5" Nov 24 17:42:54 crc kubenswrapper[4808]: I1124 17:42:54.993106 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.002152 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6f6bb48fbc-hcvb2"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.004067 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.010075 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6f6bb48fbc-hcvb2"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.021546 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e98950ba-e946-4422-8ecb-e3e5d85b4438-log-httpd\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.021639 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.021672 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-scripts\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.021717 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e98950ba-e946-4422-8ecb-e3e5d85b4438-run-httpd\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.021755 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.021789 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dht4q\" (UniqueName: \"kubernetes.io/projected/e98950ba-e946-4422-8ecb-e3e5d85b4438-kube-api-access-dht4q\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.021807 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-config-data\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.063609 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.068009 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.070422 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.070834 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xzt98" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.071000 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.071165 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.073127 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-wjrlj"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.074492 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.075990 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-4kjnz" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.078184 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.078359 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.096466 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-5cl9j"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.109502 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.117727 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-wjrlj"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.125325 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-vnqqn"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.126525 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-vnqqn" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.128426 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-k4l78" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.128639 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.128735 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-config-data\") pod \"horizon-6f6bb48fbc-hcvb2\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.128810 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-scripts\") pod \"horizon-6f6bb48fbc-hcvb2\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.128877 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dht4q\" (UniqueName: \"kubernetes.io/projected/e98950ba-e946-4422-8ecb-e3e5d85b4438-kube-api-access-dht4q\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.128936 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-config-data\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.129026 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e98950ba-e946-4422-8ecb-e3e5d85b4438-log-httpd\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.129129 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-horizon-secret-key\") pod \"horizon-6f6bb48fbc-hcvb2\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.129192 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.129272 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-scripts\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.129356 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-logs\") pod \"horizon-6f6bb48fbc-hcvb2\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.129415 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e98950ba-e946-4422-8ecb-e3e5d85b4438-run-httpd\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.129484 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcq9d\" (UniqueName: \"kubernetes.io/projected/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-kube-api-access-jcq9d\") pod \"horizon-6f6bb48fbc-hcvb2\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.129927 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.130615 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e98950ba-e946-4422-8ecb-e3e5d85b4438-log-httpd\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.133634 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e98950ba-e946-4422-8ecb-e3e5d85b4438-run-httpd\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.133993 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-p77zx" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.135666 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-config-data\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.136367 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-scripts\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.138300 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.138628 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.147997 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-4f2w2"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.150093 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kp6v5" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.150590 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.150759 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dht4q\" (UniqueName: \"kubernetes.io/projected/e98950ba-e946-4422-8ecb-e3e5d85b4438-kube-api-access-dht4q\") pod \"ceilometer-0\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.155070 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-vnqqn"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.187222 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-4f2w2"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.192765 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.199699 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.201137 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.207229 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.207298 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.217377 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233648 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-logs\") pod \"placement-db-sync-wjrlj\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233694 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233733 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233757 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59jmn\" (UniqueName: \"kubernetes.io/projected/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-kube-api-access-59jmn\") pod \"placement-db-sync-wjrlj\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233782 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-logs\") pod \"horizon-6f6bb48fbc-hcvb2\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233804 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70cb7cc9-4412-4087-860c-fd0036efe290-logs\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233823 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcq9d\" (UniqueName: \"kubernetes.io/projected/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-kube-api-access-jcq9d\") pod \"horizon-6f6bb48fbc-hcvb2\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233839 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj8cx\" (UniqueName: \"kubernetes.io/projected/70cb7cc9-4412-4087-860c-fd0036efe290-kube-api-access-xj8cx\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233865 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/70cb7cc9-4412-4087-860c-fd0036efe290-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233884 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-scripts\") pod \"placement-db-sync-wjrlj\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233898 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-config-data\") pod \"horizon-6f6bb48fbc-hcvb2\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233916 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-config-data\") pod \"placement-db-sync-wjrlj\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233937 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-scripts\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233953 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-combined-ca-bundle\") pod \"placement-db-sync-wjrlj\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233969 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-scripts\") pod \"horizon-6f6bb48fbc-hcvb2\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.233986 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.234008 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5vjz\" (UniqueName: \"kubernetes.io/projected/9ee3f387-059b-4ca5-bd55-488b737a0c12-kube-api-access-b5vjz\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.234054 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71142c75-1dcb-4b06-92bd-f9882aa6a929-combined-ca-bundle\") pod \"barbican-db-sync-vnqqn\" (UID: \"71142c75-1dcb-4b06-92bd-f9882aa6a929\") " pod="openstack/barbican-db-sync-vnqqn" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.234075 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.234088 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.234106 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-config-data\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.234123 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shhg4\" (UniqueName: \"kubernetes.io/projected/71142c75-1dcb-4b06-92bd-f9882aa6a929-kube-api-access-shhg4\") pod \"barbican-db-sync-vnqqn\" (UID: \"71142c75-1dcb-4b06-92bd-f9882aa6a929\") " pod="openstack/barbican-db-sync-vnqqn" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.234157 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/71142c75-1dcb-4b06-92bd-f9882aa6a929-db-sync-config-data\") pod \"barbican-db-sync-vnqqn\" (UID: \"71142c75-1dcb-4b06-92bd-f9882aa6a929\") " pod="openstack/barbican-db-sync-vnqqn" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.234182 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.234201 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-horizon-secret-key\") pod \"horizon-6f6bb48fbc-hcvb2\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.234225 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-config\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.234253 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.234856 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-logs\") pod \"horizon-6f6bb48fbc-hcvb2\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.235995 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-scripts\") pod \"horizon-6f6bb48fbc-hcvb2\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.237781 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-config-data\") pod \"horizon-6f6bb48fbc-hcvb2\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.244141 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-horizon-secret-key\") pod \"horizon-6f6bb48fbc-hcvb2\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.264847 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcq9d\" (UniqueName: \"kubernetes.io/projected/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-kube-api-access-jcq9d\") pod \"horizon-6f6bb48fbc-hcvb2\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.295368 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.324296 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337486 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337554 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59jmn\" (UniqueName: \"kubernetes.io/projected/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-kube-api-access-59jmn\") pod \"placement-db-sync-wjrlj\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337590 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337626 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70cb7cc9-4412-4087-860c-fd0036efe290-logs\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337655 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj8cx\" (UniqueName: \"kubernetes.io/projected/70cb7cc9-4412-4087-860c-fd0036efe290-kube-api-access-xj8cx\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337679 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab7d7aaa-4417-43c7-a604-6d70afcbec81-logs\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337698 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ab7d7aaa-4417-43c7-a604-6d70afcbec81-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337713 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/70cb7cc9-4412-4087-860c-fd0036efe290-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337736 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-scripts\") pod \"placement-db-sync-wjrlj\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337756 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-config-data\") pod \"placement-db-sync-wjrlj\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337777 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-scripts\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337794 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-combined-ca-bundle\") pod \"placement-db-sync-wjrlj\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337813 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337838 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5vjz\" (UniqueName: \"kubernetes.io/projected/9ee3f387-059b-4ca5-bd55-488b737a0c12-kube-api-access-b5vjz\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337863 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337882 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71142c75-1dcb-4b06-92bd-f9882aa6a929-combined-ca-bundle\") pod \"barbican-db-sync-vnqqn\" (UID: \"71142c75-1dcb-4b06-92bd-f9882aa6a929\") " pod="openstack/barbican-db-sync-vnqqn" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337903 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337922 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k4vx\" (UniqueName: \"kubernetes.io/projected/ab7d7aaa-4417-43c7-a604-6d70afcbec81-kube-api-access-8k4vx\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337939 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337958 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337977 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-config-data\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.337995 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shhg4\" (UniqueName: \"kubernetes.io/projected/71142c75-1dcb-4b06-92bd-f9882aa6a929-kube-api-access-shhg4\") pod \"barbican-db-sync-vnqqn\" (UID: \"71142c75-1dcb-4b06-92bd-f9882aa6a929\") " pod="openstack/barbican-db-sync-vnqqn" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.338054 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.338078 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/71142c75-1dcb-4b06-92bd-f9882aa6a929-db-sync-config-data\") pod \"barbican-db-sync-vnqqn\" (UID: \"71142c75-1dcb-4b06-92bd-f9882aa6a929\") " pod="openstack/barbican-db-sync-vnqqn" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.338097 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.338124 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.338154 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-config\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.338187 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.338202 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-logs\") pod \"placement-db-sync-wjrlj\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.338221 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.338465 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bfff8"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.338622 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.341914 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.342593 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70cb7cc9-4412-4087-860c-fd0036efe290-logs\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.343811 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/70cb7cc9-4412-4087-860c-fd0036efe290-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.353980 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-scripts\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.358219 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.358544 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.359410 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71142c75-1dcb-4b06-92bd-f9882aa6a929-combined-ca-bundle\") pod \"barbican-db-sync-vnqqn\" (UID: \"71142c75-1dcb-4b06-92bd-f9882aa6a929\") " pod="openstack/barbican-db-sync-vnqqn" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.360093 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.361564 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-config\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.362070 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-logs\") pod \"placement-db-sync-wjrlj\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.362632 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.362954 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-scripts\") pod \"placement-db-sync-wjrlj\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.363708 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-combined-ca-bundle\") pod \"placement-db-sync-wjrlj\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.371161 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-config-data\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.372291 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.377160 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/71142c75-1dcb-4b06-92bd-f9882aa6a929-db-sync-config-data\") pod \"barbican-db-sync-vnqqn\" (UID: \"71142c75-1dcb-4b06-92bd-f9882aa6a929\") " pod="openstack/barbican-db-sync-vnqqn" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.378260 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.383969 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-config-data\") pod \"placement-db-sync-wjrlj\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.397068 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59jmn\" (UniqueName: \"kubernetes.io/projected/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-kube-api-access-59jmn\") pod \"placement-db-sync-wjrlj\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.403478 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.407733 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shhg4\" (UniqueName: \"kubernetes.io/projected/71142c75-1dcb-4b06-92bd-f9882aa6a929-kube-api-access-shhg4\") pod \"barbican-db-sync-vnqqn\" (UID: \"71142c75-1dcb-4b06-92bd-f9882aa6a929\") " pod="openstack/barbican-db-sync-vnqqn" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.414917 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj8cx\" (UniqueName: \"kubernetes.io/projected/70cb7cc9-4412-4087-860c-fd0036efe290-kube-api-access-xj8cx\") pod \"glance-default-external-api-0\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.418407 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5vjz\" (UniqueName: \"kubernetes.io/projected/9ee3f387-059b-4ca5-bd55-488b737a0c12-kube-api-access-b5vjz\") pod \"dnsmasq-dns-785d8bcb8c-4f2w2\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.423210 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wjrlj" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.441107 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-5cl9j"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.442237 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-ovsdbserver-sb\") pod \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.442328 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-ovsdbserver-nb\") pod \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.442369 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-config\") pod \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.442426 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl92t\" (UniqueName: \"kubernetes.io/projected/328bde2b-812b-4cf8-b8a4-438ccbf6d108-kube-api-access-zl92t\") pod \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.442605 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-dns-svc\") pod \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\" (UID: \"328bde2b-812b-4cf8-b8a4-438ccbf6d108\") " Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.442918 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.443058 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.443095 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k4vx\" (UniqueName: \"kubernetes.io/projected/ab7d7aaa-4417-43c7-a604-6d70afcbec81-kube-api-access-8k4vx\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.443149 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.443183 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.443277 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.443331 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab7d7aaa-4417-43c7-a604-6d70afcbec81-logs\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.443353 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ab7d7aaa-4417-43c7-a604-6d70afcbec81-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.443859 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ab7d7aaa-4417-43c7-a604-6d70afcbec81-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.443903 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.445775 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-vnqqn" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.446453 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab7d7aaa-4417-43c7-a604-6d70afcbec81-logs\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.447362 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.450217 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.462570 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/328bde2b-812b-4cf8-b8a4-438ccbf6d108-kube-api-access-zl92t" (OuterVolumeSpecName: "kube-api-access-zl92t") pod "328bde2b-812b-4cf8-b8a4-438ccbf6d108" (UID: "328bde2b-812b-4cf8-b8a4-438ccbf6d108"). InnerVolumeSpecName "kube-api-access-zl92t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.475772 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.487381 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.502275 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k4vx\" (UniqueName: \"kubernetes.io/projected/ab7d7aaa-4417-43c7-a604-6d70afcbec81-kube-api-access-8k4vx\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.523238 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.558646 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl92t\" (UniqueName: \"kubernetes.io/projected/328bde2b-812b-4cf8-b8a4-438ccbf6d108-kube-api-access-zl92t\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.609450 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.702889 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.710129 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "328bde2b-812b-4cf8-b8a4-438ccbf6d108" (UID: "328bde2b-812b-4cf8-b8a4-438ccbf6d108"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.792386 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.824824 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kp6v5"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.835975 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-config" (OuterVolumeSpecName: "config") pod "328bde2b-812b-4cf8-b8a4-438ccbf6d108" (UID: "328bde2b-812b-4cf8-b8a4-438ccbf6d108"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.856186 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.870510 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f886bf6d5-mrvf5"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.894577 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.916590 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "328bde2b-812b-4cf8-b8a4-438ccbf6d108" (UID: "328bde2b-812b-4cf8-b8a4-438ccbf6d108"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.943736 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "328bde2b-812b-4cf8-b8a4-438ccbf6d108" (UID: "328bde2b-812b-4cf8-b8a4-438ccbf6d108"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.963422 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-p77zx"] Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.996926 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:55 crc kubenswrapper[4808]: I1124 17:42:55.997128 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/328bde2b-812b-4cf8-b8a4-438ccbf6d108-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.225750 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bfff8" event={"ID":"7dff84dc-26ca-4e8f-8386-37abe2041d7f","Type":"ContainerStarted","Data":"6c77be6e7a27482ef1fe9100dd64eb94eefb26e3e0196ac05e79538c5600664e"} Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.243666 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f886bf6d5-mrvf5" event={"ID":"e378c94e-0dd2-482a-bb40-6a965270a5a3","Type":"ContainerStarted","Data":"cbf8c33acef72f33ddc1c342a140e99fa2a15fda2cff1cefb242604d1b3a5852"} Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.262306 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" event={"ID":"19a0a600-f3b4-4c6f-90bb-2889ba076149","Type":"ContainerStarted","Data":"246d0dcd5fd955de493d5ba5c5ee180fae0f12b53ccb5b822e318367051feed2"} Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.268055 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kp6v5" event={"ID":"b2112bdd-0811-4bd0-927d-277b44b5cfc0","Type":"ContainerStarted","Data":"634201869afaeb2c28c5ff085a585872681ba63052cd5fa1f81b0eb5b54a8000"} Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.278030 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-f2jqs" event={"ID":"328bde2b-812b-4cf8-b8a4-438ccbf6d108","Type":"ContainerDied","Data":"79bff3c094f43a4e9f8f738eed6f321e084ffe6d2fbe9f2858928c51ff47d722"} Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.278091 4808 scope.go:117] "RemoveContainer" containerID="4b54deb9d465b1902512df3a773a8a3e978f9e9fbf6ab948b077fb78427ec67a" Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.278487 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-f2jqs" Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.314327 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-p77zx" event={"ID":"b80d86dd-fbd8-444a-8159-d5c870eb01be","Type":"ContainerStarted","Data":"1f8cffbcce837fd726f5b5a10a96d244de563377c20124de1a8e71fa836218ba"} Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.345083 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-f2jqs"] Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.350620 4808 scope.go:117] "RemoveContainer" containerID="3ab0c0a9c8b54d6e2baff9ec27547e6367217d6c746abdf098fc7a77e022c931" Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.372358 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-f2jqs"] Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.439398 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6f6bb48fbc-hcvb2"] Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.461503 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.554030 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-vnqqn"] Nov 24 17:42:56 crc kubenswrapper[4808]: W1124 17:42:56.564704 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71142c75_1dcb_4b06_92bd_f9882aa6a929.slice/crio-ea11f82ecb37819ef4c9484584cb0a4ec99cabc8a732fd458b27894260a93eb0 WatchSource:0}: Error finding container ea11f82ecb37819ef4c9484584cb0a4ec99cabc8a732fd458b27894260a93eb0: Status 404 returned error can't find the container with id ea11f82ecb37819ef4c9484584cb0a4ec99cabc8a732fd458b27894260a93eb0 Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.586611 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-wjrlj"] Nov 24 17:42:56 crc kubenswrapper[4808]: W1124 17:42:56.610195 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67ce2c95_e3aa_4d5f_8597_55c293b9e22b.slice/crio-6ae460faf1498698ecc56c3c930433f0cd32173d52be88be4a3b708758aed156 WatchSource:0}: Error finding container 6ae460faf1498698ecc56c3c930433f0cd32173d52be88be4a3b708758aed156: Status 404 returned error can't find the container with id 6ae460faf1498698ecc56c3c930433f0cd32173d52be88be4a3b708758aed156 Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.782734 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-4f2w2"] Nov 24 17:42:56 crc kubenswrapper[4808]: I1124 17:42:56.888926 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:42:56 crc kubenswrapper[4808]: W1124 17:42:56.895231 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab7d7aaa_4417_43c7_a604_6d70afcbec81.slice/crio-1c22a2efd28ba972486f31d20af59ce8e0fa3a5946775f1295b65b8851f207ec WatchSource:0}: Error finding container 1c22a2efd28ba972486f31d20af59ce8e0fa3a5946775f1295b65b8851f207ec: Status 404 returned error can't find the container with id 1c22a2efd28ba972486f31d20af59ce8e0fa3a5946775f1295b65b8851f207ec Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.351344 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kp6v5" event={"ID":"b2112bdd-0811-4bd0-927d-277b44b5cfc0","Type":"ContainerStarted","Data":"0f300ffe4c22dc83f9d9f75dc05b4ddfb6d512f92c83fd1be7ea1ee284a8a765"} Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.364335 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-vnqqn" event={"ID":"71142c75-1dcb-4b06-92bd-f9882aa6a929","Type":"ContainerStarted","Data":"ea11f82ecb37819ef4c9484584cb0a4ec99cabc8a732fd458b27894260a93eb0"} Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.365825 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ab7d7aaa-4417-43c7-a604-6d70afcbec81","Type":"ContainerStarted","Data":"1c22a2efd28ba972486f31d20af59ce8e0fa3a5946775f1295b65b8851f207ec"} Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.366941 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6f6bb48fbc-hcvb2" event={"ID":"eac7041d-3e76-4bf2-b9cd-953cc67aac6e","Type":"ContainerStarted","Data":"89971f58ea19d6b7f180fa317d5545da359b250926186b0ecae3cd23b3e28810"} Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.369205 4808 generic.go:334] "Generic (PLEG): container finished" podID="9ee3f387-059b-4ca5-bd55-488b737a0c12" containerID="2ae865cf786b6e42f1815bc5e71fb46309b6c2b051521e8390781bc858facd93" exitCode=0 Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.369249 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" event={"ID":"9ee3f387-059b-4ca5-bd55-488b737a0c12","Type":"ContainerDied","Data":"2ae865cf786b6e42f1815bc5e71fb46309b6c2b051521e8390781bc858facd93"} Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.369265 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" event={"ID":"9ee3f387-059b-4ca5-bd55-488b737a0c12","Type":"ContainerStarted","Data":"2bd382437b8b61dddc6964224b17d1d5af836d0c98787d5e6d1230761623f574"} Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.380392 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-kp6v5" podStartSLOduration=3.380376794 podStartE2EDuration="3.380376794s" podCreationTimestamp="2025-11-24 17:42:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:42:57.37785519 +0000 UTC m=+969.975522992" watchObservedRunningTime="2025-11-24 17:42:57.380376794 +0000 UTC m=+969.978044596" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.381390 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bfff8" event={"ID":"7dff84dc-26ca-4e8f-8386-37abe2041d7f","Type":"ContainerStarted","Data":"fbc61c0dd82b553d021d24e75615ff8ba915ee6ff26a60b71262cac688eabbcf"} Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.387358 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wjrlj" event={"ID":"67ce2c95-e3aa-4d5f-8597-55c293b9e22b","Type":"ContainerStarted","Data":"6ae460faf1498698ecc56c3c930433f0cd32173d52be88be4a3b708758aed156"} Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.429344 4808 generic.go:334] "Generic (PLEG): container finished" podID="19a0a600-f3b4-4c6f-90bb-2889ba076149" containerID="12ee4e0ca921e8dff2e6354c6d52cdf90b668a19e0e7514a03b5a20c6f73041d" exitCode=0 Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.429458 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" event={"ID":"19a0a600-f3b4-4c6f-90bb-2889ba076149","Type":"ContainerDied","Data":"12ee4e0ca921e8dff2e6354c6d52cdf90b668a19e0e7514a03b5a20c6f73041d"} Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.449954 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e98950ba-e946-4422-8ecb-e3e5d85b4438","Type":"ContainerStarted","Data":"07368dcfe91bf73acafeb65b73c9a3e60b7b7f3e9e3ddbced1a27c4b97d753ca"} Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.522751 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-bfff8" podStartSLOduration=3.522731435 podStartE2EDuration="3.522731435s" podCreationTimestamp="2025-11-24 17:42:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:42:57.442916503 +0000 UTC m=+970.040584305" watchObservedRunningTime="2025-11-24 17:42:57.522731435 +0000 UTC m=+970.120399237" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.594282 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6f6bb48fbc-hcvb2"] Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.609194 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.636746 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-645c99cbf7-9qwvk"] Nov 24 17:42:57 crc kubenswrapper[4808]: E1124 17:42:57.637153 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="328bde2b-812b-4cf8-b8a4-438ccbf6d108" containerName="init" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.637165 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="328bde2b-812b-4cf8-b8a4-438ccbf6d108" containerName="init" Nov 24 17:42:57 crc kubenswrapper[4808]: E1124 17:42:57.637176 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="328bde2b-812b-4cf8-b8a4-438ccbf6d108" containerName="dnsmasq-dns" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.637181 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="328bde2b-812b-4cf8-b8a4-438ccbf6d108" containerName="dnsmasq-dns" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.637391 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="328bde2b-812b-4cf8-b8a4-438ccbf6d108" containerName="dnsmasq-dns" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.638316 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.679663 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-645c99cbf7-9qwvk"] Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.712882 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.726845 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.769703 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8999\" (UniqueName: \"kubernetes.io/projected/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-kube-api-access-b8999\") pod \"horizon-645c99cbf7-9qwvk\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.770105 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-scripts\") pod \"horizon-645c99cbf7-9qwvk\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.770425 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-horizon-secret-key\") pod \"horizon-645c99cbf7-9qwvk\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.770630 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-config-data\") pod \"horizon-645c99cbf7-9qwvk\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.770678 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-logs\") pod \"horizon-645c99cbf7-9qwvk\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.797196 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.872742 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-scripts\") pod \"horizon-645c99cbf7-9qwvk\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.872820 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-horizon-secret-key\") pod \"horizon-645c99cbf7-9qwvk\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.872910 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-config-data\") pod \"horizon-645c99cbf7-9qwvk\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.872927 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-logs\") pod \"horizon-645c99cbf7-9qwvk\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.872999 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8999\" (UniqueName: \"kubernetes.io/projected/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-kube-api-access-b8999\") pod \"horizon-645c99cbf7-9qwvk\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.875058 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-scripts\") pod \"horizon-645c99cbf7-9qwvk\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.875394 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-logs\") pod \"horizon-645c99cbf7-9qwvk\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.876226 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-config-data\") pod \"horizon-645c99cbf7-9qwvk\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.878423 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-horizon-secret-key\") pod \"horizon-645c99cbf7-9qwvk\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.891175 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8999\" (UniqueName: \"kubernetes.io/projected/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-kube-api-access-b8999\") pod \"horizon-645c99cbf7-9qwvk\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:57 crc kubenswrapper[4808]: I1124 17:42:57.987600 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.129141 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.282406 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-config\") pod \"19a0a600-f3b4-4c6f-90bb-2889ba076149\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.282536 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw82n\" (UniqueName: \"kubernetes.io/projected/19a0a600-f3b4-4c6f-90bb-2889ba076149-kube-api-access-zw82n\") pod \"19a0a600-f3b4-4c6f-90bb-2889ba076149\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.282667 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-ovsdbserver-nb\") pod \"19a0a600-f3b4-4c6f-90bb-2889ba076149\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.282849 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-dns-svc\") pod \"19a0a600-f3b4-4c6f-90bb-2889ba076149\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.282898 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-ovsdbserver-sb\") pod \"19a0a600-f3b4-4c6f-90bb-2889ba076149\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.282922 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-dns-swift-storage-0\") pod \"19a0a600-f3b4-4c6f-90bb-2889ba076149\" (UID: \"19a0a600-f3b4-4c6f-90bb-2889ba076149\") " Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.305416 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19a0a600-f3b4-4c6f-90bb-2889ba076149-kube-api-access-zw82n" (OuterVolumeSpecName: "kube-api-access-zw82n") pod "19a0a600-f3b4-4c6f-90bb-2889ba076149" (UID: "19a0a600-f3b4-4c6f-90bb-2889ba076149"). InnerVolumeSpecName "kube-api-access-zw82n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.329842 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "19a0a600-f3b4-4c6f-90bb-2889ba076149" (UID: "19a0a600-f3b4-4c6f-90bb-2889ba076149"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.332730 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "19a0a600-f3b4-4c6f-90bb-2889ba076149" (UID: "19a0a600-f3b4-4c6f-90bb-2889ba076149"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.340685 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-config" (OuterVolumeSpecName: "config") pod "19a0a600-f3b4-4c6f-90bb-2889ba076149" (UID: "19a0a600-f3b4-4c6f-90bb-2889ba076149"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.345476 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "19a0a600-f3b4-4c6f-90bb-2889ba076149" (UID: "19a0a600-f3b4-4c6f-90bb-2889ba076149"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.354338 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "19a0a600-f3b4-4c6f-90bb-2889ba076149" (UID: "19a0a600-f3b4-4c6f-90bb-2889ba076149"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.386988 4808 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.387065 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.387081 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw82n\" (UniqueName: \"kubernetes.io/projected/19a0a600-f3b4-4c6f-90bb-2889ba076149-kube-api-access-zw82n\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.387094 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.387107 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.387148 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19a0a600-f3b4-4c6f-90bb-2889ba076149-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.391702 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="328bde2b-812b-4cf8-b8a4-438ccbf6d108" path="/var/lib/kubelet/pods/328bde2b-812b-4cf8-b8a4-438ccbf6d108/volumes" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.569646 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ab7d7aaa-4417-43c7-a604-6d70afcbec81","Type":"ContainerStarted","Data":"f0609da4df0063dad0bc753bf7fd2a1998537327ec2279eaab6237f3247eada1"} Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.582074 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-645c99cbf7-9qwvk"] Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.604211 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"70cb7cc9-4412-4087-860c-fd0036efe290","Type":"ContainerStarted","Data":"fe0ab8490f650514f2fcee72fbcd509682b8dd65b3b32fa2a1e341656274aeeb"} Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.621571 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" event={"ID":"19a0a600-f3b4-4c6f-90bb-2889ba076149","Type":"ContainerDied","Data":"246d0dcd5fd955de493d5ba5c5ee180fae0f12b53ccb5b822e318367051feed2"} Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.621625 4808 scope.go:117] "RemoveContainer" containerID="12ee4e0ca921e8dff2e6354c6d52cdf90b668a19e0e7514a03b5a20c6f73041d" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.621754 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-5cl9j" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.655689 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" event={"ID":"9ee3f387-059b-4ca5-bd55-488b737a0c12","Type":"ContainerStarted","Data":"9bd29801d8ef589fdf3b93006e8a0418eafd525e3f324cd7540d1c30ee458604"} Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.655731 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.735404 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-5cl9j"] Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.737447 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-5cl9j"] Nov 24 17:42:58 crc kubenswrapper[4808]: W1124 17:42:58.744425 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f7556a8_bb8a_4c77_8775_81f0f24fbb8a.slice/crio-71448619363de9c8b6de36c913ce6bc776553f6e82610744fb428faf29a2897c WatchSource:0}: Error finding container 71448619363de9c8b6de36c913ce6bc776553f6e82610744fb428faf29a2897c: Status 404 returned error can't find the container with id 71448619363de9c8b6de36c913ce6bc776553f6e82610744fb428faf29a2897c Nov 24 17:42:58 crc kubenswrapper[4808]: I1124 17:42:58.749500 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" podStartSLOduration=4.749482356 podStartE2EDuration="4.749482356s" podCreationTimestamp="2025-11-24 17:42:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:42:58.74929809 +0000 UTC m=+971.346965892" watchObservedRunningTime="2025-11-24 17:42:58.749482356 +0000 UTC m=+971.347150158" Nov 24 17:42:59 crc kubenswrapper[4808]: I1124 17:42:59.667988 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-645c99cbf7-9qwvk" event={"ID":"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a","Type":"ContainerStarted","Data":"71448619363de9c8b6de36c913ce6bc776553f6e82610744fb428faf29a2897c"} Nov 24 17:42:59 crc kubenswrapper[4808]: I1124 17:42:59.672061 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ab7d7aaa-4417-43c7-a604-6d70afcbec81","Type":"ContainerStarted","Data":"2c177be9d8ea5934739f2c37dbe6013329b0de44325d6137cc9197d23eeab062"} Nov 24 17:43:00 crc kubenswrapper[4808]: I1124 17:43:00.366691 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19a0a600-f3b4-4c6f-90bb-2889ba076149" path="/var/lib/kubelet/pods/19a0a600-f3b4-4c6f-90bb-2889ba076149/volumes" Nov 24 17:43:00 crc kubenswrapper[4808]: I1124 17:43:00.693692 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ab7d7aaa-4417-43c7-a604-6d70afcbec81" containerName="glance-log" containerID="cri-o://f0609da4df0063dad0bc753bf7fd2a1998537327ec2279eaab6237f3247eada1" gracePeriod=30 Nov 24 17:43:00 crc kubenswrapper[4808]: I1124 17:43:00.693975 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"70cb7cc9-4412-4087-860c-fd0036efe290","Type":"ContainerStarted","Data":"32c0bfa9c0061673aa127648b626f2be200b1aa7e153b2a634b2645e26ddb829"} Nov 24 17:43:00 crc kubenswrapper[4808]: I1124 17:43:00.694256 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ab7d7aaa-4417-43c7-a604-6d70afcbec81" containerName="glance-httpd" containerID="cri-o://2c177be9d8ea5934739f2c37dbe6013329b0de44325d6137cc9197d23eeab062" gracePeriod=30 Nov 24 17:43:00 crc kubenswrapper[4808]: I1124 17:43:00.729928 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.729911163 podStartE2EDuration="6.729911163s" podCreationTimestamp="2025-11-24 17:42:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:43:00.722788696 +0000 UTC m=+973.320456498" watchObservedRunningTime="2025-11-24 17:43:00.729911163 +0000 UTC m=+973.327578965" Nov 24 17:43:01 crc kubenswrapper[4808]: I1124 17:43:01.718468 4808 generic.go:334] "Generic (PLEG): container finished" podID="ab7d7aaa-4417-43c7-a604-6d70afcbec81" containerID="2c177be9d8ea5934739f2c37dbe6013329b0de44325d6137cc9197d23eeab062" exitCode=0 Nov 24 17:43:01 crc kubenswrapper[4808]: I1124 17:43:01.718852 4808 generic.go:334] "Generic (PLEG): container finished" podID="ab7d7aaa-4417-43c7-a604-6d70afcbec81" containerID="f0609da4df0063dad0bc753bf7fd2a1998537327ec2279eaab6237f3247eada1" exitCode=143 Nov 24 17:43:01 crc kubenswrapper[4808]: I1124 17:43:01.718745 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ab7d7aaa-4417-43c7-a604-6d70afcbec81","Type":"ContainerDied","Data":"2c177be9d8ea5934739f2c37dbe6013329b0de44325d6137cc9197d23eeab062"} Nov 24 17:43:01 crc kubenswrapper[4808]: I1124 17:43:01.718892 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ab7d7aaa-4417-43c7-a604-6d70afcbec81","Type":"ContainerDied","Data":"f0609da4df0063dad0bc753bf7fd2a1998537327ec2279eaab6237f3247eada1"} Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.038966 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-f886bf6d5-mrvf5"] Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.062674 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-74bc8cd64-b5zpt"] Nov 24 17:43:04 crc kubenswrapper[4808]: E1124 17:43:04.063199 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19a0a600-f3b4-4c6f-90bb-2889ba076149" containerName="init" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.063245 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="19a0a600-f3b4-4c6f-90bb-2889ba076149" containerName="init" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.063465 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="19a0a600-f3b4-4c6f-90bb-2889ba076149" containerName="init" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.064520 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.070862 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.087755 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74bc8cd64-b5zpt"] Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.123116 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pbbs\" (UniqueName: \"kubernetes.io/projected/da69b7d2-5a35-4dd7-accb-937fefd0d94f-kube-api-access-7pbbs\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.123202 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-horizon-secret-key\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.123270 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da69b7d2-5a35-4dd7-accb-937fefd0d94f-scripts\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.123297 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-horizon-tls-certs\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.123329 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da69b7d2-5a35-4dd7-accb-937fefd0d94f-logs\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.123468 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da69b7d2-5a35-4dd7-accb-937fefd0d94f-config-data\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.123514 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-combined-ca-bundle\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.173673 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-645c99cbf7-9qwvk"] Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.209583 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5ddbdb6df8-pv5zj"] Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.211075 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.217261 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5ddbdb6df8-pv5zj"] Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.225950 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da69b7d2-5a35-4dd7-accb-937fefd0d94f-config-data\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.226037 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-combined-ca-bundle\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.226096 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pbbs\" (UniqueName: \"kubernetes.io/projected/da69b7d2-5a35-4dd7-accb-937fefd0d94f-kube-api-access-7pbbs\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.226152 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-horizon-secret-key\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.226209 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da69b7d2-5a35-4dd7-accb-937fefd0d94f-scripts\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.226234 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-horizon-tls-certs\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.226263 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da69b7d2-5a35-4dd7-accb-937fefd0d94f-logs\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.226768 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da69b7d2-5a35-4dd7-accb-937fefd0d94f-logs\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.228005 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da69b7d2-5a35-4dd7-accb-937fefd0d94f-config-data\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.232688 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da69b7d2-5a35-4dd7-accb-937fefd0d94f-scripts\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.238467 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-horizon-secret-key\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.240646 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-combined-ca-bundle\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.261094 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-horizon-tls-certs\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.275542 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pbbs\" (UniqueName: \"kubernetes.io/projected/da69b7d2-5a35-4dd7-accb-937fefd0d94f-kube-api-access-7pbbs\") pod \"horizon-74bc8cd64-b5zpt\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.327726 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/96142f0a-3231-4617-bbd2-9f440c73908e-horizon-tls-certs\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.327772 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/96142f0a-3231-4617-bbd2-9f440c73908e-horizon-secret-key\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.327801 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96142f0a-3231-4617-bbd2-9f440c73908e-logs\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.327824 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2x5q\" (UniqueName: \"kubernetes.io/projected/96142f0a-3231-4617-bbd2-9f440c73908e-kube-api-access-z2x5q\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.327969 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96142f0a-3231-4617-bbd2-9f440c73908e-config-data\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.327990 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96142f0a-3231-4617-bbd2-9f440c73908e-scripts\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.328081 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96142f0a-3231-4617-bbd2-9f440c73908e-combined-ca-bundle\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.391839 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.429920 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/96142f0a-3231-4617-bbd2-9f440c73908e-horizon-tls-certs\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.430292 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/96142f0a-3231-4617-bbd2-9f440c73908e-horizon-secret-key\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.430345 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96142f0a-3231-4617-bbd2-9f440c73908e-logs\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.430370 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2x5q\" (UniqueName: \"kubernetes.io/projected/96142f0a-3231-4617-bbd2-9f440c73908e-kube-api-access-z2x5q\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.430445 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96142f0a-3231-4617-bbd2-9f440c73908e-config-data\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.430466 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96142f0a-3231-4617-bbd2-9f440c73908e-scripts\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.430532 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96142f0a-3231-4617-bbd2-9f440c73908e-combined-ca-bundle\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.432592 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96142f0a-3231-4617-bbd2-9f440c73908e-logs\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.433156 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96142f0a-3231-4617-bbd2-9f440c73908e-config-data\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.433721 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96142f0a-3231-4617-bbd2-9f440c73908e-scripts\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.436839 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96142f0a-3231-4617-bbd2-9f440c73908e-combined-ca-bundle\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.436992 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/96142f0a-3231-4617-bbd2-9f440c73908e-horizon-tls-certs\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.439985 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/96142f0a-3231-4617-bbd2-9f440c73908e-horizon-secret-key\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.453111 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2x5q\" (UniqueName: \"kubernetes.io/projected/96142f0a-3231-4617-bbd2-9f440c73908e-kube-api-access-z2x5q\") pod \"horizon-5ddbdb6df8-pv5zj\" (UID: \"96142f0a-3231-4617-bbd2-9f440c73908e\") " pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.532816 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:04 crc kubenswrapper[4808]: I1124 17:43:04.860064 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74bc8cd64-b5zpt"] Nov 24 17:43:04 crc kubenswrapper[4808]: W1124 17:43:04.875775 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda69b7d2_5a35_4dd7_accb_937fefd0d94f.slice/crio-9ae975cae99004f67fe3732de7bf04f15dab40f283df331982a18682869c14a9 WatchSource:0}: Error finding container 9ae975cae99004f67fe3732de7bf04f15dab40f283df331982a18682869c14a9: Status 404 returned error can't find the container with id 9ae975cae99004f67fe3732de7bf04f15dab40f283df331982a18682869c14a9 Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.003929 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5ddbdb6df8-pv5zj"] Nov 24 17:43:05 crc kubenswrapper[4808]: W1124 17:43:05.013256 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96142f0a_3231_4617_bbd2_9f440c73908e.slice/crio-4e5335c00861b918502ae33bf0ed93aa154388a1e1c45aee0038b0e93bb28966 WatchSource:0}: Error finding container 4e5335c00861b918502ae33bf0ed93aa154388a1e1c45aee0038b0e93bb28966: Status 404 returned error can't find the container with id 4e5335c00861b918502ae33bf0ed93aa154388a1e1c45aee0038b0e93bb28966 Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.533220 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.609246 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-xhk99"] Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.609510 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" podUID="45350630-8ea8-4ca6-83b1-f5a8cb26afb9" containerName="dnsmasq-dns" containerID="cri-o://726ba42ada2711d0602c876e9abe812da88a44225ec61712c953d52f5c328479" gracePeriod=10 Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.631853 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.676936 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab7d7aaa-4417-43c7-a604-6d70afcbec81-logs\") pod \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.677112 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ab7d7aaa-4417-43c7-a604-6d70afcbec81-httpd-run\") pod \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.677168 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-internal-tls-certs\") pod \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.677265 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-combined-ca-bundle\") pod \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.677366 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-config-data\") pod \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.677413 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8k4vx\" (UniqueName: \"kubernetes.io/projected/ab7d7aaa-4417-43c7-a604-6d70afcbec81-kube-api-access-8k4vx\") pod \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.677509 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-scripts\") pod \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.677541 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\" (UID: \"ab7d7aaa-4417-43c7-a604-6d70afcbec81\") " Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.678809 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab7d7aaa-4417-43c7-a604-6d70afcbec81-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ab7d7aaa-4417-43c7-a604-6d70afcbec81" (UID: "ab7d7aaa-4417-43c7-a604-6d70afcbec81"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.679241 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab7d7aaa-4417-43c7-a604-6d70afcbec81-logs" (OuterVolumeSpecName: "logs") pod "ab7d7aaa-4417-43c7-a604-6d70afcbec81" (UID: "ab7d7aaa-4417-43c7-a604-6d70afcbec81"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.701076 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "ab7d7aaa-4417-43c7-a604-6d70afcbec81" (UID: "ab7d7aaa-4417-43c7-a604-6d70afcbec81"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.706164 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab7d7aaa-4417-43c7-a604-6d70afcbec81-kube-api-access-8k4vx" (OuterVolumeSpecName: "kube-api-access-8k4vx") pod "ab7d7aaa-4417-43c7-a604-6d70afcbec81" (UID: "ab7d7aaa-4417-43c7-a604-6d70afcbec81"). InnerVolumeSpecName "kube-api-access-8k4vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.709288 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-scripts" (OuterVolumeSpecName: "scripts") pod "ab7d7aaa-4417-43c7-a604-6d70afcbec81" (UID: "ab7d7aaa-4417-43c7-a604-6d70afcbec81"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.762750 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab7d7aaa-4417-43c7-a604-6d70afcbec81" (UID: "ab7d7aaa-4417-43c7-a604-6d70afcbec81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.765972 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74bc8cd64-b5zpt" event={"ID":"da69b7d2-5a35-4dd7-accb-937fefd0d94f","Type":"ContainerStarted","Data":"9ae975cae99004f67fe3732de7bf04f15dab40f283df331982a18682869c14a9"} Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.766398 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ab7d7aaa-4417-43c7-a604-6d70afcbec81" (UID: "ab7d7aaa-4417-43c7-a604-6d70afcbec81"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.775243 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ab7d7aaa-4417-43c7-a604-6d70afcbec81","Type":"ContainerDied","Data":"1c22a2efd28ba972486f31d20af59ce8e0fa3a5946775f1295b65b8851f207ec"} Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.775959 4808 scope.go:117] "RemoveContainer" containerID="2c177be9d8ea5934739f2c37dbe6013329b0de44325d6137cc9197d23eeab062" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.776258 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.780862 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-config-data" (OuterVolumeSpecName: "config-data") pod "ab7d7aaa-4417-43c7-a604-6d70afcbec81" (UID: "ab7d7aaa-4417-43c7-a604-6d70afcbec81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.782222 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.782245 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.782254 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8k4vx\" (UniqueName: \"kubernetes.io/projected/ab7d7aaa-4417-43c7-a604-6d70afcbec81-kube-api-access-8k4vx\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.782265 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.782290 4808 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.782300 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab7d7aaa-4417-43c7-a604-6d70afcbec81-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.782308 4808 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ab7d7aaa-4417-43c7-a604-6d70afcbec81-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.782316 4808 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab7d7aaa-4417-43c7-a604-6d70afcbec81-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.801352 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"70cb7cc9-4412-4087-860c-fd0036efe290","Type":"ContainerStarted","Data":"2208adae172aec26a99fd710d0a2421e8e73db47280552a077a077d2323d0f8a"} Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.801515 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="70cb7cc9-4412-4087-860c-fd0036efe290" containerName="glance-log" containerID="cri-o://32c0bfa9c0061673aa127648b626f2be200b1aa7e153b2a634b2645e26ddb829" gracePeriod=30 Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.801987 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="70cb7cc9-4412-4087-860c-fd0036efe290" containerName="glance-httpd" containerID="cri-o://2208adae172aec26a99fd710d0a2421e8e73db47280552a077a077d2323d0f8a" gracePeriod=30 Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.807820 4808 generic.go:334] "Generic (PLEG): container finished" podID="45350630-8ea8-4ca6-83b1-f5a8cb26afb9" containerID="726ba42ada2711d0602c876e9abe812da88a44225ec61712c953d52f5c328479" exitCode=0 Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.807883 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" event={"ID":"45350630-8ea8-4ca6-83b1-f5a8cb26afb9","Type":"ContainerDied","Data":"726ba42ada2711d0602c876e9abe812da88a44225ec61712c953d52f5c328479"} Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.809991 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5ddbdb6df8-pv5zj" event={"ID":"96142f0a-3231-4617-bbd2-9f440c73908e","Type":"ContainerStarted","Data":"4e5335c00861b918502ae33bf0ed93aa154388a1e1c45aee0038b0e93bb28966"} Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.824216 4808 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.839629 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.839336072 podStartE2EDuration="11.839336072s" podCreationTimestamp="2025-11-24 17:42:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:43:05.826437725 +0000 UTC m=+978.424105527" watchObservedRunningTime="2025-11-24 17:43:05.839336072 +0000 UTC m=+978.437003874" Nov 24 17:43:05 crc kubenswrapper[4808]: I1124 17:43:05.884171 4808 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.115806 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.130655 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.146209 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:43:06 crc kubenswrapper[4808]: E1124 17:43:06.146593 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab7d7aaa-4417-43c7-a604-6d70afcbec81" containerName="glance-log" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.146617 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab7d7aaa-4417-43c7-a604-6d70afcbec81" containerName="glance-log" Nov 24 17:43:06 crc kubenswrapper[4808]: E1124 17:43:06.146634 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab7d7aaa-4417-43c7-a604-6d70afcbec81" containerName="glance-httpd" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.146641 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab7d7aaa-4417-43c7-a604-6d70afcbec81" containerName="glance-httpd" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.147593 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab7d7aaa-4417-43c7-a604-6d70afcbec81" containerName="glance-httpd" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.147634 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab7d7aaa-4417-43c7-a604-6d70afcbec81" containerName="glance-log" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.178988 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.180900 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.188124 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.189216 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.308347 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21208d77-22ac-428a-a5a0-39c69fe1fa6f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.308695 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.308771 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.308862 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21208d77-22ac-428a-a5a0-39c69fe1fa6f-logs\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.308933 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.309034 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.309115 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58wdw\" (UniqueName: \"kubernetes.io/projected/21208d77-22ac-428a-a5a0-39c69fe1fa6f-kube-api-access-58wdw\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.309192 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.359141 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab7d7aaa-4417-43c7-a604-6d70afcbec81" path="/var/lib/kubelet/pods/ab7d7aaa-4417-43c7-a604-6d70afcbec81/volumes" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.410639 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21208d77-22ac-428a-a5a0-39c69fe1fa6f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.410802 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.410827 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.410884 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21208d77-22ac-428a-a5a0-39c69fe1fa6f-logs\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.410920 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.410972 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.410994 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58wdw\" (UniqueName: \"kubernetes.io/projected/21208d77-22ac-428a-a5a0-39c69fe1fa6f-kube-api-access-58wdw\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.411046 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.411413 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.412150 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21208d77-22ac-428a-a5a0-39c69fe1fa6f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.412531 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21208d77-22ac-428a-a5a0-39c69fe1fa6f-logs\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.418133 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.418982 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.419675 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.437845 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.438470 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58wdw\" (UniqueName: \"kubernetes.io/projected/21208d77-22ac-428a-a5a0-39c69fe1fa6f-kube-api-access-58wdw\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.441381 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.501895 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.523432 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.523494 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.825850 4808 generic.go:334] "Generic (PLEG): container finished" podID="70cb7cc9-4412-4087-860c-fd0036efe290" containerID="2208adae172aec26a99fd710d0a2421e8e73db47280552a077a077d2323d0f8a" exitCode=0 Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.825882 4808 generic.go:334] "Generic (PLEG): container finished" podID="70cb7cc9-4412-4087-860c-fd0036efe290" containerID="32c0bfa9c0061673aa127648b626f2be200b1aa7e153b2a634b2645e26ddb829" exitCode=143 Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.825887 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"70cb7cc9-4412-4087-860c-fd0036efe290","Type":"ContainerDied","Data":"2208adae172aec26a99fd710d0a2421e8e73db47280552a077a077d2323d0f8a"} Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.825919 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"70cb7cc9-4412-4087-860c-fd0036efe290","Type":"ContainerDied","Data":"32c0bfa9c0061673aa127648b626f2be200b1aa7e153b2a634b2645e26ddb829"} Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.827956 4808 generic.go:334] "Generic (PLEG): container finished" podID="7dff84dc-26ca-4e8f-8386-37abe2041d7f" containerID="fbc61c0dd82b553d021d24e75615ff8ba915ee6ff26a60b71262cac688eabbcf" exitCode=0 Nov 24 17:43:06 crc kubenswrapper[4808]: I1124 17:43:06.827978 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bfff8" event={"ID":"7dff84dc-26ca-4e8f-8386-37abe2041d7f","Type":"ContainerDied","Data":"fbc61c0dd82b553d021d24e75615ff8ba915ee6ff26a60b71262cac688eabbcf"} Nov 24 17:43:08 crc kubenswrapper[4808]: I1124 17:43:08.779036 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" podUID="45350630-8ea8-4ca6-83b1-f5a8cb26afb9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.137:5353: connect: connection refused" Nov 24 17:43:16 crc kubenswrapper[4808]: E1124 17:43:16.189619 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 24 17:43:16 crc kubenswrapper[4808]: E1124 17:43:16.190335 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h6fh644h8h57bhdbh56bh5fh5dch5f7h65dh558h599h8dh5bfh647h67bh88h5fdh5f7h55h5d7h67fh557h58h697h576h54bh5bch65fh57bh5dcq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-527cr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-f886bf6d5-mrvf5_openstack(e378c94e-0dd2-482a-bb40-6a965270a5a3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:43:16 crc kubenswrapper[4808]: E1124 17:43:16.194837 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-f886bf6d5-mrvf5" podUID="e378c94e-0dd2-482a-bb40-6a965270a5a3" Nov 24 17:43:18 crc kubenswrapper[4808]: E1124 17:43:18.056658 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Nov 24 17:43:18 crc kubenswrapper[4808]: E1124 17:43:18.057450 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-59jmn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-wjrlj_openstack(67ce2c95-e3aa-4d5f-8597-55c293b9e22b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:43:18 crc kubenswrapper[4808]: E1124 17:43:18.058722 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-wjrlj" podUID="67ce2c95-e3aa-4d5f-8597-55c293b9e22b" Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.137199 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:43:18 crc kubenswrapper[4808]: E1124 17:43:18.164241 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 24 17:43:18 crc kubenswrapper[4808]: E1124 17:43:18.164443 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5fch5c6h57ch696h5cch565h588h5f9h86h6h5f4h557h59chd6h655h5bbhfdh589hdbh5bhbh685h9fh5hc4h644h5b7hc7h64bh668h584hdfq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jcq9d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6f6bb48fbc-hcvb2_openstack(eac7041d-3e76-4bf2-b9cd-953cc67aac6e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:43:18 crc kubenswrapper[4808]: E1124 17:43:18.167371 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-6f6bb48fbc-hcvb2" podUID="eac7041d-3e76-4bf2-b9cd-953cc67aac6e" Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.333714 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e378c94e-0dd2-482a-bb40-6a965270a5a3-config-data\") pod \"e378c94e-0dd2-482a-bb40-6a965270a5a3\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.333794 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e378c94e-0dd2-482a-bb40-6a965270a5a3-scripts\") pod \"e378c94e-0dd2-482a-bb40-6a965270a5a3\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.333887 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-527cr\" (UniqueName: \"kubernetes.io/projected/e378c94e-0dd2-482a-bb40-6a965270a5a3-kube-api-access-527cr\") pod \"e378c94e-0dd2-482a-bb40-6a965270a5a3\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.333996 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e378c94e-0dd2-482a-bb40-6a965270a5a3-horizon-secret-key\") pod \"e378c94e-0dd2-482a-bb40-6a965270a5a3\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.334061 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e378c94e-0dd2-482a-bb40-6a965270a5a3-logs\") pod \"e378c94e-0dd2-482a-bb40-6a965270a5a3\" (UID: \"e378c94e-0dd2-482a-bb40-6a965270a5a3\") " Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.334977 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e378c94e-0dd2-482a-bb40-6a965270a5a3-logs" (OuterVolumeSpecName: "logs") pod "e378c94e-0dd2-482a-bb40-6a965270a5a3" (UID: "e378c94e-0dd2-482a-bb40-6a965270a5a3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.344109 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e378c94e-0dd2-482a-bb40-6a965270a5a3-kube-api-access-527cr" (OuterVolumeSpecName: "kube-api-access-527cr") pod "e378c94e-0dd2-482a-bb40-6a965270a5a3" (UID: "e378c94e-0dd2-482a-bb40-6a965270a5a3"). InnerVolumeSpecName "kube-api-access-527cr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.344721 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e378c94e-0dd2-482a-bb40-6a965270a5a3-scripts" (OuterVolumeSpecName: "scripts") pod "e378c94e-0dd2-482a-bb40-6a965270a5a3" (UID: "e378c94e-0dd2-482a-bb40-6a965270a5a3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.358726 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e378c94e-0dd2-482a-bb40-6a965270a5a3-config-data" (OuterVolumeSpecName: "config-data") pod "e378c94e-0dd2-482a-bb40-6a965270a5a3" (UID: "e378c94e-0dd2-482a-bb40-6a965270a5a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.368345 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e378c94e-0dd2-482a-bb40-6a965270a5a3-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e378c94e-0dd2-482a-bb40-6a965270a5a3" (UID: "e378c94e-0dd2-482a-bb40-6a965270a5a3"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.436150 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-527cr\" (UniqueName: \"kubernetes.io/projected/e378c94e-0dd2-482a-bb40-6a965270a5a3-kube-api-access-527cr\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.436195 4808 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e378c94e-0dd2-482a-bb40-6a965270a5a3-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.436207 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e378c94e-0dd2-482a-bb40-6a965270a5a3-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.436219 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e378c94e-0dd2-482a-bb40-6a965270a5a3-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.436231 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e378c94e-0dd2-482a-bb40-6a965270a5a3-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.780051 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" podUID="45350630-8ea8-4ca6-83b1-f5a8cb26afb9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.137:5353: i/o timeout" Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.933845 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f886bf6d5-mrvf5" Nov 24 17:43:18 crc kubenswrapper[4808]: I1124 17:43:18.934711 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f886bf6d5-mrvf5" event={"ID":"e378c94e-0dd2-482a-bb40-6a965270a5a3","Type":"ContainerDied","Data":"cbf8c33acef72f33ddc1c342a140e99fa2a15fda2cff1cefb242604d1b3a5852"} Nov 24 17:43:18 crc kubenswrapper[4808]: E1124 17:43:18.938743 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-wjrlj" podUID="67ce2c95-e3aa-4d5f-8597-55c293b9e22b" Nov 24 17:43:19 crc kubenswrapper[4808]: I1124 17:43:19.060086 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-f886bf6d5-mrvf5"] Nov 24 17:43:19 crc kubenswrapper[4808]: I1124 17:43:19.066778 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-f886bf6d5-mrvf5"] Nov 24 17:43:20 crc kubenswrapper[4808]: I1124 17:43:20.357203 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e378c94e-0dd2-482a-bb40-6a965270a5a3" path="/var/lib/kubelet/pods/e378c94e-0dd2-482a-bb40-6a965270a5a3/volumes" Nov 24 17:43:23 crc kubenswrapper[4808]: I1124 17:43:23.781256 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" podUID="45350630-8ea8-4ca6-83b1-f5a8cb26afb9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.137:5353: i/o timeout" Nov 24 17:43:23 crc kubenswrapper[4808]: I1124 17:43:23.782053 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:43:25 crc kubenswrapper[4808]: I1124 17:43:25.703424 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 17:43:25 crc kubenswrapper[4808]: I1124 17:43:25.703896 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 17:43:28 crc kubenswrapper[4808]: I1124 17:43:28.782338 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" podUID="45350630-8ea8-4ca6-83b1-f5a8cb26afb9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.137:5353: i/o timeout" Nov 24 17:43:29 crc kubenswrapper[4808]: E1124 17:43:29.136664 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 24 17:43:29 crc kubenswrapper[4808]: E1124 17:43:29.136876 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n577h598h5dbh5c8h4h5ffh664h678h666h65fh548h89h5dfh669hbh554h5c7h648h64fh54fh97h67dh57bhcbhf5h567h5f5hb7h89hc8h5d7h5dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7pbbs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-74bc8cd64-b5zpt_openstack(da69b7d2-5a35-4dd7-accb-937fefd0d94f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:43:29 crc kubenswrapper[4808]: E1124 17:43:29.140357 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-74bc8cd64-b5zpt" podUID="da69b7d2-5a35-4dd7-accb-937fefd0d94f" Nov 24 17:43:29 crc kubenswrapper[4808]: E1124 17:43:29.156292 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 24 17:43:29 crc kubenswrapper[4808]: E1124 17:43:29.156522 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n677h58bh657hcfh5b7h7bh584h5cch6fh65bh594hb5h99h5b8h659h54h54ch9bh64h54fh64dh699h5fch5c5h84h54fh85h58dhc4h664h5b6h55fq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b8999,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-645c99cbf7-9qwvk_openstack(8f7556a8-bb8a-4c77-8775-81f0f24fbb8a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:43:29 crc kubenswrapper[4808]: E1124 17:43:29.159867 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-645c99cbf7-9qwvk" podUID="8f7556a8-bb8a-4c77-8775-81f0f24fbb8a" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.260005 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.443806 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-scripts\") pod \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.443879 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-fernet-keys\") pod \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.443980 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-config-data\") pod \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.444063 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clm7p\" (UniqueName: \"kubernetes.io/projected/7dff84dc-26ca-4e8f-8386-37abe2041d7f-kube-api-access-clm7p\") pod \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.444097 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-combined-ca-bundle\") pod \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.444120 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-credential-keys\") pod \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\" (UID: \"7dff84dc-26ca-4e8f-8386-37abe2041d7f\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.450533 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-scripts" (OuterVolumeSpecName: "scripts") pod "7dff84dc-26ca-4e8f-8386-37abe2041d7f" (UID: "7dff84dc-26ca-4e8f-8386-37abe2041d7f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.451082 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7dff84dc-26ca-4e8f-8386-37abe2041d7f" (UID: "7dff84dc-26ca-4e8f-8386-37abe2041d7f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.451164 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dff84dc-26ca-4e8f-8386-37abe2041d7f-kube-api-access-clm7p" (OuterVolumeSpecName: "kube-api-access-clm7p") pod "7dff84dc-26ca-4e8f-8386-37abe2041d7f" (UID: "7dff84dc-26ca-4e8f-8386-37abe2041d7f"). InnerVolumeSpecName "kube-api-access-clm7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.463448 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "7dff84dc-26ca-4e8f-8386-37abe2041d7f" (UID: "7dff84dc-26ca-4e8f-8386-37abe2041d7f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.474788 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7dff84dc-26ca-4e8f-8386-37abe2041d7f" (UID: "7dff84dc-26ca-4e8f-8386-37abe2041d7f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.477527 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-config-data" (OuterVolumeSpecName: "config-data") pod "7dff84dc-26ca-4e8f-8386-37abe2041d7f" (UID: "7dff84dc-26ca-4e8f-8386-37abe2041d7f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.546296 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.546339 4808 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.546349 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.546358 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clm7p\" (UniqueName: \"kubernetes.io/projected/7dff84dc-26ca-4e8f-8386-37abe2041d7f-kube-api-access-clm7p\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.546369 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.546377 4808 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7dff84dc-26ca-4e8f-8386-37abe2041d7f-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:29 crc kubenswrapper[4808]: E1124 17:43:29.709338 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 24 17:43:29 crc kubenswrapper[4808]: E1124 17:43:29.709549 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-shhg4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-vnqqn_openstack(71142c75-1dcb-4b06-92bd-f9882aa6a929): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.709827 4808 scope.go:117] "RemoveContainer" containerID="f0609da4df0063dad0bc753bf7fd2a1998537327ec2279eaab6237f3247eada1" Nov 24 17:43:29 crc kubenswrapper[4808]: E1124 17:43:29.710682 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-vnqqn" podUID="71142c75-1dcb-4b06-92bd-f9882aa6a929" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.746550 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.753250 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.797173 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.854532 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-ovsdbserver-sb\") pod \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.854623 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-dns-swift-storage-0\") pod \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.854674 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjh4b\" (UniqueName: \"kubernetes.io/projected/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-kube-api-access-pjh4b\") pod \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.854748 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-dns-svc\") pod \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.854825 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-ovsdbserver-nb\") pod \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.854859 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-config\") pod \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\" (UID: \"45350630-8ea8-4ca6-83b1-f5a8cb26afb9\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.862302 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-kube-api-access-pjh4b" (OuterVolumeSpecName: "kube-api-access-pjh4b") pod "45350630-8ea8-4ca6-83b1-f5a8cb26afb9" (UID: "45350630-8ea8-4ca6-83b1-f5a8cb26afb9"). InnerVolumeSpecName "kube-api-access-pjh4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.907341 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "45350630-8ea8-4ca6-83b1-f5a8cb26afb9" (UID: "45350630-8ea8-4ca6-83b1-f5a8cb26afb9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.910235 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "45350630-8ea8-4ca6-83b1-f5a8cb26afb9" (UID: "45350630-8ea8-4ca6-83b1-f5a8cb26afb9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.914586 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "45350630-8ea8-4ca6-83b1-f5a8cb26afb9" (UID: "45350630-8ea8-4ca6-83b1-f5a8cb26afb9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.928760 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "45350630-8ea8-4ca6-83b1-f5a8cb26afb9" (UID: "45350630-8ea8-4ca6-83b1-f5a8cb26afb9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.932422 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-config" (OuterVolumeSpecName: "config") pod "45350630-8ea8-4ca6-83b1-f5a8cb26afb9" (UID: "45350630-8ea8-4ca6-83b1-f5a8cb26afb9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.956351 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-logs\") pod \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.956433 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-public-tls-certs\") pod \"70cb7cc9-4412-4087-860c-fd0036efe290\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.956510 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-scripts\") pod \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.956561 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xj8cx\" (UniqueName: \"kubernetes.io/projected/70cb7cc9-4412-4087-860c-fd0036efe290-kube-api-access-xj8cx\") pod \"70cb7cc9-4412-4087-860c-fd0036efe290\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.956803 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-logs" (OuterVolumeSpecName: "logs") pod "eac7041d-3e76-4bf2-b9cd-953cc67aac6e" (UID: "eac7041d-3e76-4bf2-b9cd-953cc67aac6e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.957450 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-scripts" (OuterVolumeSpecName: "scripts") pod "eac7041d-3e76-4bf2-b9cd-953cc67aac6e" (UID: "eac7041d-3e76-4bf2-b9cd-953cc67aac6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.957529 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"70cb7cc9-4412-4087-860c-fd0036efe290\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.957557 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcq9d\" (UniqueName: \"kubernetes.io/projected/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-kube-api-access-jcq9d\") pod \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.957806 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-combined-ca-bundle\") pod \"70cb7cc9-4412-4087-860c-fd0036efe290\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.957838 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-scripts\") pod \"70cb7cc9-4412-4087-860c-fd0036efe290\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.957856 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-horizon-secret-key\") pod \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.957899 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70cb7cc9-4412-4087-860c-fd0036efe290-logs\") pod \"70cb7cc9-4412-4087-860c-fd0036efe290\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.957930 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-config-data\") pod \"70cb7cc9-4412-4087-860c-fd0036efe290\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.958036 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-config-data\") pod \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\" (UID: \"eac7041d-3e76-4bf2-b9cd-953cc67aac6e\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.958067 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/70cb7cc9-4412-4087-860c-fd0036efe290-httpd-run\") pod \"70cb7cc9-4412-4087-860c-fd0036efe290\" (UID: \"70cb7cc9-4412-4087-860c-fd0036efe290\") " Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.958403 4808 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.958416 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjh4b\" (UniqueName: \"kubernetes.io/projected/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-kube-api-access-pjh4b\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.958427 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.958435 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.958443 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.958452 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.958463 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.958474 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/45350630-8ea8-4ca6-83b1-f5a8cb26afb9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.958669 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70cb7cc9-4412-4087-860c-fd0036efe290-logs" (OuterVolumeSpecName: "logs") pod "70cb7cc9-4412-4087-860c-fd0036efe290" (UID: "70cb7cc9-4412-4087-860c-fd0036efe290"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.959491 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70cb7cc9-4412-4087-860c-fd0036efe290-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "70cb7cc9-4412-4087-860c-fd0036efe290" (UID: "70cb7cc9-4412-4087-860c-fd0036efe290"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.959731 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-config-data" (OuterVolumeSpecName: "config-data") pod "eac7041d-3e76-4bf2-b9cd-953cc67aac6e" (UID: "eac7041d-3e76-4bf2-b9cd-953cc67aac6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.962854 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-kube-api-access-jcq9d" (OuterVolumeSpecName: "kube-api-access-jcq9d") pod "eac7041d-3e76-4bf2-b9cd-953cc67aac6e" (UID: "eac7041d-3e76-4bf2-b9cd-953cc67aac6e"). InnerVolumeSpecName "kube-api-access-jcq9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.963242 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-scripts" (OuterVolumeSpecName: "scripts") pod "70cb7cc9-4412-4087-860c-fd0036efe290" (UID: "70cb7cc9-4412-4087-860c-fd0036efe290"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.963396 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "70cb7cc9-4412-4087-860c-fd0036efe290" (UID: "70cb7cc9-4412-4087-860c-fd0036efe290"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.963765 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "eac7041d-3e76-4bf2-b9cd-953cc67aac6e" (UID: "eac7041d-3e76-4bf2-b9cd-953cc67aac6e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.964924 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70cb7cc9-4412-4087-860c-fd0036efe290-kube-api-access-xj8cx" (OuterVolumeSpecName: "kube-api-access-xj8cx") pod "70cb7cc9-4412-4087-860c-fd0036efe290" (UID: "70cb7cc9-4412-4087-860c-fd0036efe290"). InnerVolumeSpecName "kube-api-access-xj8cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:29 crc kubenswrapper[4808]: I1124 17:43:29.989553 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70cb7cc9-4412-4087-860c-fd0036efe290" (UID: "70cb7cc9-4412-4087-860c-fd0036efe290"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.007577 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-config-data" (OuterVolumeSpecName: "config-data") pod "70cb7cc9-4412-4087-860c-fd0036efe290" (UID: "70cb7cc9-4412-4087-860c-fd0036efe290"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.030487 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "70cb7cc9-4412-4087-860c-fd0036efe290" (UID: "70cb7cc9-4412-4087-860c-fd0036efe290"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.034753 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"70cb7cc9-4412-4087-860c-fd0036efe290","Type":"ContainerDied","Data":"fe0ab8490f650514f2fcee72fbcd509682b8dd65b3b32fa2a1e341656274aeeb"} Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.034820 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.036341 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6f6bb48fbc-hcvb2" event={"ID":"eac7041d-3e76-4bf2-b9cd-953cc67aac6e","Type":"ContainerDied","Data":"89971f58ea19d6b7f180fa317d5545da359b250926186b0ecae3cd23b3e28810"} Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.036448 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6f6bb48fbc-hcvb2" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.043235 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" event={"ID":"45350630-8ea8-4ca6-83b1-f5a8cb26afb9","Type":"ContainerDied","Data":"cfb38c13e871e09c2000a534524cd6dc138bad9fd5d921ad0f6e5ef92269aed3"} Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.043381 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.046074 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bfff8" event={"ID":"7dff84dc-26ca-4e8f-8386-37abe2041d7f","Type":"ContainerDied","Data":"6c77be6e7a27482ef1fe9100dd64eb94eefb26e3e0196ac05e79538c5600664e"} Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.046134 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c77be6e7a27482ef1fe9100dd64eb94eefb26e3e0196ac05e79538c5600664e" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.046161 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bfff8" Nov 24 17:43:30 crc kubenswrapper[4808]: E1124 17:43:30.048136 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-vnqqn" podUID="71142c75-1dcb-4b06-92bd-f9882aa6a929" Nov 24 17:43:30 crc kubenswrapper[4808]: E1124 17:43:30.048803 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-74bc8cd64-b5zpt" podUID="da69b7d2-5a35-4dd7-accb-937fefd0d94f" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.069135 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70cb7cc9-4412-4087-860c-fd0036efe290-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.069172 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.069187 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.069202 4808 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/70cb7cc9-4412-4087-860c-fd0036efe290-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.069214 4808 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.069229 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xj8cx\" (UniqueName: \"kubernetes.io/projected/70cb7cc9-4412-4087-860c-fd0036efe290-kube-api-access-xj8cx\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.069258 4808 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.069272 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcq9d\" (UniqueName: \"kubernetes.io/projected/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-kube-api-access-jcq9d\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.069285 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.069297 4808 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eac7041d-3e76-4bf2-b9cd-953cc67aac6e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.069309 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70cb7cc9-4412-4087-860c-fd0036efe290-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.100713 4808 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.170561 4808 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.185834 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6f6bb48fbc-hcvb2"] Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.212103 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6f6bb48fbc-hcvb2"] Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.218493 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.224840 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.232237 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-xhk99"] Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.238082 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-xhk99"] Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.246262 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:43:30 crc kubenswrapper[4808]: E1124 17:43:30.246815 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70cb7cc9-4412-4087-860c-fd0036efe290" containerName="glance-httpd" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.246841 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="70cb7cc9-4412-4087-860c-fd0036efe290" containerName="glance-httpd" Nov 24 17:43:30 crc kubenswrapper[4808]: E1124 17:43:30.246873 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45350630-8ea8-4ca6-83b1-f5a8cb26afb9" containerName="init" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.246880 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="45350630-8ea8-4ca6-83b1-f5a8cb26afb9" containerName="init" Nov 24 17:43:30 crc kubenswrapper[4808]: E1124 17:43:30.246890 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dff84dc-26ca-4e8f-8386-37abe2041d7f" containerName="keystone-bootstrap" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.246898 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dff84dc-26ca-4e8f-8386-37abe2041d7f" containerName="keystone-bootstrap" Nov 24 17:43:30 crc kubenswrapper[4808]: E1124 17:43:30.246913 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45350630-8ea8-4ca6-83b1-f5a8cb26afb9" containerName="dnsmasq-dns" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.246919 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="45350630-8ea8-4ca6-83b1-f5a8cb26afb9" containerName="dnsmasq-dns" Nov 24 17:43:30 crc kubenswrapper[4808]: E1124 17:43:30.246933 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70cb7cc9-4412-4087-860c-fd0036efe290" containerName="glance-log" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.246938 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="70cb7cc9-4412-4087-860c-fd0036efe290" containerName="glance-log" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.247188 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dff84dc-26ca-4e8f-8386-37abe2041d7f" containerName="keystone-bootstrap" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.247210 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="70cb7cc9-4412-4087-860c-fd0036efe290" containerName="glance-httpd" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.247228 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="45350630-8ea8-4ca6-83b1-f5a8cb26afb9" containerName="dnsmasq-dns" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.247252 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="70cb7cc9-4412-4087-860c-fd0036efe290" containerName="glance-log" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.249495 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.253083 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.253202 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.263837 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.358844 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45350630-8ea8-4ca6-83b1-f5a8cb26afb9" path="/var/lib/kubelet/pods/45350630-8ea8-4ca6-83b1-f5a8cb26afb9/volumes" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.360056 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70cb7cc9-4412-4087-860c-fd0036efe290" path="/var/lib/kubelet/pods/70cb7cc9-4412-4087-860c-fd0036efe290/volumes" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.360804 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eac7041d-3e76-4bf2-b9cd-953cc67aac6e" path="/var/lib/kubelet/pods/eac7041d-3e76-4bf2-b9cd-953cc67aac6e/volumes" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.375503 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-config-data\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.375570 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjzbr\" (UniqueName: \"kubernetes.io/projected/25fd4604-12b2-412f-8d76-15584feda527-kube-api-access-gjzbr\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.375639 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.375712 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-scripts\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.375756 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25fd4604-12b2-412f-8d76-15584feda527-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.375777 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.375801 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25fd4604-12b2-412f-8d76-15584feda527-logs\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.375873 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.398999 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-bfff8"] Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.406554 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-bfff8"] Nov 24 17:43:30 crc kubenswrapper[4808]: E1124 17:43:30.428944 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Nov 24 17:43:30 crc kubenswrapper[4808]: E1124 17:43:30.429170 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n64hcdh6ch559h5c7h599h678h55ch79hbfh64dh55bh674h57bh599h557h5b5h655h555h66dh578h59h58dh5h688h544h5ch585h68dhd5hfh88q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dht4q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(e98950ba-e946-4422-8ecb-e3e5d85b4438): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.477407 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-scripts\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.477559 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25fd4604-12b2-412f-8d76-15584feda527-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.477589 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.477613 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25fd4604-12b2-412f-8d76-15584feda527-logs\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.477759 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.477836 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-config-data\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.477904 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjzbr\" (UniqueName: \"kubernetes.io/projected/25fd4604-12b2-412f-8d76-15584feda527-kube-api-access-gjzbr\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.478076 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.478744 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25fd4604-12b2-412f-8d76-15584feda527-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.479273 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25fd4604-12b2-412f-8d76-15584feda527-logs\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.479761 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.484355 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-scripts\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.484567 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.485567 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.498756 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-config-data\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.503887 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjzbr\" (UniqueName: \"kubernetes.io/projected/25fd4604-12b2-412f-8d76-15584feda527-kube-api-access-gjzbr\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.506645 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dj9zb"] Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.507970 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.512406 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.512705 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.513573 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.513855 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.518941 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-622rb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.518593 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dj9zb"] Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.529301 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.582369 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.681316 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-fernet-keys\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.681368 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-scripts\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.681404 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-config-data\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.681492 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-credential-keys\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.681558 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x66cr\" (UniqueName: \"kubernetes.io/projected/56775e00-b320-4260-b162-e99c948df631-kube-api-access-x66cr\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.681939 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-combined-ca-bundle\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.783777 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-fernet-keys\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.785067 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-scripts\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.785112 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-config-data\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.785169 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-credential-keys\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.785227 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x66cr\" (UniqueName: \"kubernetes.io/projected/56775e00-b320-4260-b162-e99c948df631-kube-api-access-x66cr\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.785394 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-combined-ca-bundle\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.789520 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-fernet-keys\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.790094 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-scripts\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.790300 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-credential-keys\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.790579 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-config-data\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.790617 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-combined-ca-bundle\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.806574 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x66cr\" (UniqueName: \"kubernetes.io/projected/56775e00-b320-4260-b162-e99c948df631-kube-api-access-x66cr\") pod \"keystone-bootstrap-dj9zb\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:30 crc kubenswrapper[4808]: I1124 17:43:30.895712 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.508428 4808 scope.go:117] "RemoveContainer" containerID="2208adae172aec26a99fd710d0a2421e8e73db47280552a077a077d2323d0f8a" Nov 24 17:43:31 crc kubenswrapper[4808]: E1124 17:43:31.548213 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 24 17:43:31 crc kubenswrapper[4808]: E1124 17:43:31.548386 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4l74k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-p77zx_openstack(b80d86dd-fbd8-444a-8159-d5c870eb01be): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:43:31 crc kubenswrapper[4808]: E1124 17:43:31.549568 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-p77zx" podUID="b80d86dd-fbd8-444a-8159-d5c870eb01be" Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.636810 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.643143 4808 scope.go:117] "RemoveContainer" containerID="32c0bfa9c0061673aa127648b626f2be200b1aa7e153b2a634b2645e26ddb829" Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.702029 4808 scope.go:117] "RemoveContainer" containerID="726ba42ada2711d0602c876e9abe812da88a44225ec61712c953d52f5c328479" Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.729510 4808 scope.go:117] "RemoveContainer" containerID="ba881a33e0592c1120cad1782d30e5127b95e61fef296581b977002ccd71649e" Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.801632 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-horizon-secret-key\") pod \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.802051 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-logs\") pod \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.802122 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-scripts\") pod \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.802142 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-config-data\") pod \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.802253 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8999\" (UniqueName: \"kubernetes.io/projected/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-kube-api-access-b8999\") pod \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\" (UID: \"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a\") " Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.803418 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-logs" (OuterVolumeSpecName: "logs") pod "8f7556a8-bb8a-4c77-8775-81f0f24fbb8a" (UID: "8f7556a8-bb8a-4c77-8775-81f0f24fbb8a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.803840 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-scripts" (OuterVolumeSpecName: "scripts") pod "8f7556a8-bb8a-4c77-8775-81f0f24fbb8a" (UID: "8f7556a8-bb8a-4c77-8775-81f0f24fbb8a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.804125 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.804155 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.804562 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-config-data" (OuterVolumeSpecName: "config-data") pod "8f7556a8-bb8a-4c77-8775-81f0f24fbb8a" (UID: "8f7556a8-bb8a-4c77-8775-81f0f24fbb8a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.813132 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8f7556a8-bb8a-4c77-8775-81f0f24fbb8a" (UID: "8f7556a8-bb8a-4c77-8775-81f0f24fbb8a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.816240 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-kube-api-access-b8999" (OuterVolumeSpecName: "kube-api-access-b8999") pod "8f7556a8-bb8a-4c77-8775-81f0f24fbb8a" (UID: "8f7556a8-bb8a-4c77-8775-81f0f24fbb8a"). InnerVolumeSpecName "kube-api-access-b8999". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.905736 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8999\" (UniqueName: \"kubernetes.io/projected/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-kube-api-access-b8999\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.905773 4808 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:31 crc kubenswrapper[4808]: I1124 17:43:31.905783 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:32 crc kubenswrapper[4808]: I1124 17:43:32.046866 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dj9zb"] Nov 24 17:43:32 crc kubenswrapper[4808]: I1124 17:43:32.069048 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-645c99cbf7-9qwvk" event={"ID":"8f7556a8-bb8a-4c77-8775-81f0f24fbb8a","Type":"ContainerDied","Data":"71448619363de9c8b6de36c913ce6bc776553f6e82610744fb428faf29a2897c"} Nov 24 17:43:32 crc kubenswrapper[4808]: I1124 17:43:32.069077 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-645c99cbf7-9qwvk" Nov 24 17:43:32 crc kubenswrapper[4808]: I1124 17:43:32.072472 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5ddbdb6df8-pv5zj" event={"ID":"96142f0a-3231-4617-bbd2-9f440c73908e","Type":"ContainerStarted","Data":"2ae1b181f2c896f8b85c3fd0f5c4e842dfb5dba36747c3e0b9cb197bbedf4843"} Nov 24 17:43:32 crc kubenswrapper[4808]: I1124 17:43:32.072527 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5ddbdb6df8-pv5zj" event={"ID":"96142f0a-3231-4617-bbd2-9f440c73908e","Type":"ContainerStarted","Data":"c1332f65365467a9736154273961f5d934aca46fa695fc76debe1168beea24cf"} Nov 24 17:43:32 crc kubenswrapper[4808]: E1124 17:43:32.077666 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-p77zx" podUID="b80d86dd-fbd8-444a-8159-d5c870eb01be" Nov 24 17:43:32 crc kubenswrapper[4808]: I1124 17:43:32.119917 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5ddbdb6df8-pv5zj" podStartSLOduration=1.6312381070000002 podStartE2EDuration="28.119895448s" podCreationTimestamp="2025-11-24 17:43:04 +0000 UTC" firstStartedPulling="2025-11-24 17:43:05.020586661 +0000 UTC m=+977.618254453" lastFinishedPulling="2025-11-24 17:43:31.509243992 +0000 UTC m=+1004.106911794" observedRunningTime="2025-11-24 17:43:32.09771516 +0000 UTC m=+1004.695382972" watchObservedRunningTime="2025-11-24 17:43:32.119895448 +0000 UTC m=+1004.717563250" Nov 24 17:43:32 crc kubenswrapper[4808]: I1124 17:43:32.143239 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:43:32 crc kubenswrapper[4808]: I1124 17:43:32.162437 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-645c99cbf7-9qwvk"] Nov 24 17:43:32 crc kubenswrapper[4808]: I1124 17:43:32.172594 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-645c99cbf7-9qwvk"] Nov 24 17:43:32 crc kubenswrapper[4808]: I1124 17:43:32.224461 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:43:32 crc kubenswrapper[4808]: W1124 17:43:32.255008 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56775e00_b320_4260_b162_e99c948df631.slice/crio-c5f09cef0b7a041cd7bb55b965866d83a5d756ef7501e75b61f7e536a779cf36 WatchSource:0}: Error finding container c5f09cef0b7a041cd7bb55b965866d83a5d756ef7501e75b61f7e536a779cf36: Status 404 returned error can't find the container with id c5f09cef0b7a041cd7bb55b965866d83a5d756ef7501e75b61f7e536a779cf36 Nov 24 17:43:32 crc kubenswrapper[4808]: W1124 17:43:32.257957 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25fd4604_12b2_412f_8d76_15584feda527.slice/crio-684be5c0677e8b973085e0d12298e58f1d12a90cf8a17a56b199dd6bccb248fe WatchSource:0}: Error finding container 684be5c0677e8b973085e0d12298e58f1d12a90cf8a17a56b199dd6bccb248fe: Status 404 returned error can't find the container with id 684be5c0677e8b973085e0d12298e58f1d12a90cf8a17a56b199dd6bccb248fe Nov 24 17:43:32 crc kubenswrapper[4808]: W1124 17:43:32.260731 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21208d77_22ac_428a_a5a0_39c69fe1fa6f.slice/crio-d2555b21a92578f9090a7c0f0753a046ea50529b6e2084bfec5bdda50d586cae WatchSource:0}: Error finding container d2555b21a92578f9090a7c0f0753a046ea50529b6e2084bfec5bdda50d586cae: Status 404 returned error can't find the container with id d2555b21a92578f9090a7c0f0753a046ea50529b6e2084bfec5bdda50d586cae Nov 24 17:43:32 crc kubenswrapper[4808]: I1124 17:43:32.360316 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dff84dc-26ca-4e8f-8386-37abe2041d7f" path="/var/lib/kubelet/pods/7dff84dc-26ca-4e8f-8386-37abe2041d7f/volumes" Nov 24 17:43:32 crc kubenswrapper[4808]: I1124 17:43:32.361204 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f7556a8-bb8a-4c77-8775-81f0f24fbb8a" path="/var/lib/kubelet/pods/8f7556a8-bb8a-4c77-8775-81f0f24fbb8a/volumes" Nov 24 17:43:33 crc kubenswrapper[4808]: I1124 17:43:33.103295 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"25fd4604-12b2-412f-8d76-15584feda527","Type":"ContainerStarted","Data":"3f2f328d9ec9832c76186617ad9dafc44adaa3d7d110568d0dbaede229b47bc3"} Nov 24 17:43:33 crc kubenswrapper[4808]: I1124 17:43:33.103773 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"25fd4604-12b2-412f-8d76-15584feda527","Type":"ContainerStarted","Data":"684be5c0677e8b973085e0d12298e58f1d12a90cf8a17a56b199dd6bccb248fe"} Nov 24 17:43:33 crc kubenswrapper[4808]: I1124 17:43:33.107609 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e98950ba-e946-4422-8ecb-e3e5d85b4438","Type":"ContainerStarted","Data":"7c4b5968a5c94d367cfd0b20b8d880adb83a2f8780d016045aeda94ab1ed7ce9"} Nov 24 17:43:33 crc kubenswrapper[4808]: I1124 17:43:33.110821 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dj9zb" event={"ID":"56775e00-b320-4260-b162-e99c948df631","Type":"ContainerStarted","Data":"6187b82d0d321402af6fcc5c3a2a3b802f357dff62d1fc526ebb3c3fce434725"} Nov 24 17:43:33 crc kubenswrapper[4808]: I1124 17:43:33.110870 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dj9zb" event={"ID":"56775e00-b320-4260-b162-e99c948df631","Type":"ContainerStarted","Data":"c5f09cef0b7a041cd7bb55b965866d83a5d756ef7501e75b61f7e536a779cf36"} Nov 24 17:43:33 crc kubenswrapper[4808]: I1124 17:43:33.113668 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"21208d77-22ac-428a-a5a0-39c69fe1fa6f","Type":"ContainerStarted","Data":"f54aee1e3c31595f4cc7fd12294fbcf12fd01b06a02f9d9c509fe9430a81d96e"} Nov 24 17:43:33 crc kubenswrapper[4808]: I1124 17:43:33.113742 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"21208d77-22ac-428a-a5a0-39c69fe1fa6f","Type":"ContainerStarted","Data":"d2555b21a92578f9090a7c0f0753a046ea50529b6e2084bfec5bdda50d586cae"} Nov 24 17:43:33 crc kubenswrapper[4808]: I1124 17:43:33.136863 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dj9zb" podStartSLOduration=3.136833125 podStartE2EDuration="3.136833125s" podCreationTimestamp="2025-11-24 17:43:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:43:33.129062018 +0000 UTC m=+1005.726729830" watchObservedRunningTime="2025-11-24 17:43:33.136833125 +0000 UTC m=+1005.734500927" Nov 24 17:43:33 crc kubenswrapper[4808]: I1124 17:43:33.784160 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-xhk99" podUID="45350630-8ea8-4ca6-83b1-f5a8cb26afb9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.137:5353: i/o timeout" Nov 24 17:43:34 crc kubenswrapper[4808]: I1124 17:43:34.127804 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"21208d77-22ac-428a-a5a0-39c69fe1fa6f","Type":"ContainerStarted","Data":"8996b50c2d9fcdd9da3aa3e14d475f35e97b6f081cfc97b0e348e7fb1d0972c9"} Nov 24 17:43:34 crc kubenswrapper[4808]: I1124 17:43:34.132081 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"25fd4604-12b2-412f-8d76-15584feda527","Type":"ContainerStarted","Data":"67c1c320b47c34a7be66808fdc9afea31d85a153ca43935389616dd85d6cb5b0"} Nov 24 17:43:34 crc kubenswrapper[4808]: I1124 17:43:34.152457 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=28.152437243 podStartE2EDuration="28.152437243s" podCreationTimestamp="2025-11-24 17:43:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:43:34.14721017 +0000 UTC m=+1006.744877972" watchObservedRunningTime="2025-11-24 17:43:34.152437243 +0000 UTC m=+1006.750105045" Nov 24 17:43:34 crc kubenswrapper[4808]: I1124 17:43:34.180782 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.18075701 podStartE2EDuration="4.18075701s" podCreationTimestamp="2025-11-24 17:43:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:43:34.170744077 +0000 UTC m=+1006.768411899" watchObservedRunningTime="2025-11-24 17:43:34.18075701 +0000 UTC m=+1006.778424812" Nov 24 17:43:34 crc kubenswrapper[4808]: I1124 17:43:34.533166 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:34 crc kubenswrapper[4808]: I1124 17:43:34.535536 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:43:36 crc kubenswrapper[4808]: I1124 17:43:36.502844 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 17:43:36 crc kubenswrapper[4808]: I1124 17:43:36.503545 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 17:43:36 crc kubenswrapper[4808]: I1124 17:43:36.503732 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 17:43:36 crc kubenswrapper[4808]: I1124 17:43:36.503759 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 17:43:36 crc kubenswrapper[4808]: I1124 17:43:36.522599 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:43:36 crc kubenswrapper[4808]: I1124 17:43:36.522648 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:43:36 crc kubenswrapper[4808]: I1124 17:43:36.545673 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 17:43:36 crc kubenswrapper[4808]: I1124 17:43:36.553727 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 17:43:38 crc kubenswrapper[4808]: I1124 17:43:38.174508 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wjrlj" event={"ID":"67ce2c95-e3aa-4d5f-8597-55c293b9e22b","Type":"ContainerStarted","Data":"b91696c1a899298ddd07d20470b57336c1acdd027253fcf40687a2cb526de70c"} Nov 24 17:43:38 crc kubenswrapper[4808]: I1124 17:43:38.177463 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e98950ba-e946-4422-8ecb-e3e5d85b4438","Type":"ContainerStarted","Data":"9abef3dccd5ae469421834c4a60c5afa0540d004107a53962e0092a019e66cab"} Nov 24 17:43:38 crc kubenswrapper[4808]: I1124 17:43:38.179277 4808 generic.go:334] "Generic (PLEG): container finished" podID="56775e00-b320-4260-b162-e99c948df631" containerID="6187b82d0d321402af6fcc5c3a2a3b802f357dff62d1fc526ebb3c3fce434725" exitCode=0 Nov 24 17:43:38 crc kubenswrapper[4808]: I1124 17:43:38.179308 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dj9zb" event={"ID":"56775e00-b320-4260-b162-e99c948df631","Type":"ContainerDied","Data":"6187b82d0d321402af6fcc5c3a2a3b802f357dff62d1fc526ebb3c3fce434725"} Nov 24 17:43:38 crc kubenswrapper[4808]: I1124 17:43:38.191745 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-wjrlj" podStartSLOduration=2.991639541 podStartE2EDuration="44.191722238s" podCreationTimestamp="2025-11-24 17:42:54 +0000 UTC" firstStartedPulling="2025-11-24 17:42:56.617296013 +0000 UTC m=+969.214963815" lastFinishedPulling="2025-11-24 17:43:37.81737871 +0000 UTC m=+1010.415046512" observedRunningTime="2025-11-24 17:43:38.188676909 +0000 UTC m=+1010.786344721" watchObservedRunningTime="2025-11-24 17:43:38.191722238 +0000 UTC m=+1010.789390040" Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.189464 4808 generic.go:334] "Generic (PLEG): container finished" podID="b2112bdd-0811-4bd0-927d-277b44b5cfc0" containerID="0f300ffe4c22dc83f9d9f75dc05b4ddfb6d512f92c83fd1be7ea1ee284a8a765" exitCode=0 Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.189563 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kp6v5" event={"ID":"b2112bdd-0811-4bd0-927d-277b44b5cfc0","Type":"ContainerDied","Data":"0f300ffe4c22dc83f9d9f75dc05b4ddfb6d512f92c83fd1be7ea1ee284a8a765"} Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.314529 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.314913 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.551661 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.558628 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-credential-keys\") pod \"56775e00-b320-4260-b162-e99c948df631\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.558776 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-fernet-keys\") pod \"56775e00-b320-4260-b162-e99c948df631\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.558842 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x66cr\" (UniqueName: \"kubernetes.io/projected/56775e00-b320-4260-b162-e99c948df631-kube-api-access-x66cr\") pod \"56775e00-b320-4260-b162-e99c948df631\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.558897 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-scripts\") pod \"56775e00-b320-4260-b162-e99c948df631\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.558961 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-combined-ca-bundle\") pod \"56775e00-b320-4260-b162-e99c948df631\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.559104 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-config-data\") pod \"56775e00-b320-4260-b162-e99c948df631\" (UID: \"56775e00-b320-4260-b162-e99c948df631\") " Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.565545 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-scripts" (OuterVolumeSpecName: "scripts") pod "56775e00-b320-4260-b162-e99c948df631" (UID: "56775e00-b320-4260-b162-e99c948df631"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.568250 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "56775e00-b320-4260-b162-e99c948df631" (UID: "56775e00-b320-4260-b162-e99c948df631"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.575340 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "56775e00-b320-4260-b162-e99c948df631" (UID: "56775e00-b320-4260-b162-e99c948df631"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.577748 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56775e00-b320-4260-b162-e99c948df631-kube-api-access-x66cr" (OuterVolumeSpecName: "kube-api-access-x66cr") pod "56775e00-b320-4260-b162-e99c948df631" (UID: "56775e00-b320-4260-b162-e99c948df631"). InnerVolumeSpecName "kube-api-access-x66cr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.594212 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56775e00-b320-4260-b162-e99c948df631" (UID: "56775e00-b320-4260-b162-e99c948df631"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.615872 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-config-data" (OuterVolumeSpecName: "config-data") pod "56775e00-b320-4260-b162-e99c948df631" (UID: "56775e00-b320-4260-b162-e99c948df631"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.661392 4808 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.661456 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x66cr\" (UniqueName: \"kubernetes.io/projected/56775e00-b320-4260-b162-e99c948df631-kube-api-access-x66cr\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.661474 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.661486 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.661497 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:39 crc kubenswrapper[4808]: I1124 17:43:39.661507 4808 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56775e00-b320-4260-b162-e99c948df631-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.202716 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dj9zb" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.202763 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dj9zb" event={"ID":"56775e00-b320-4260-b162-e99c948df631","Type":"ContainerDied","Data":"c5f09cef0b7a041cd7bb55b965866d83a5d756ef7501e75b61f7e536a779cf36"} Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.202792 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5f09cef0b7a041cd7bb55b965866d83a5d756ef7501e75b61f7e536a779cf36" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.398729 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7b7d8889-c8cpc"] Nov 24 17:43:40 crc kubenswrapper[4808]: E1124 17:43:40.399135 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56775e00-b320-4260-b162-e99c948df631" containerName="keystone-bootstrap" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.399152 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="56775e00-b320-4260-b162-e99c948df631" containerName="keystone-bootstrap" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.399313 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="56775e00-b320-4260-b162-e99c948df631" containerName="keystone-bootstrap" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.399838 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.404247 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.404712 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-622rb" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.404762 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.404906 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.404950 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.405114 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.451218 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7b7d8889-c8cpc"] Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.476763 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-internal-tls-certs\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.476863 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-combined-ca-bundle\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.476912 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-credential-keys\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.476983 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-fernet-keys\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.477034 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-config-data\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.477051 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6tw9\" (UniqueName: \"kubernetes.io/projected/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-kube-api-access-n6tw9\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.477109 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-public-tls-certs\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.477133 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-scripts\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.580079 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-credential-keys\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.580465 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-fernet-keys\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.580513 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-config-data\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.580536 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6tw9\" (UniqueName: \"kubernetes.io/projected/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-kube-api-access-n6tw9\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.580620 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-public-tls-certs\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.580648 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-scripts\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.580694 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-internal-tls-certs\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.580727 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-combined-ca-bundle\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.582835 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.582965 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.587550 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-fernet-keys\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.588686 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-internal-tls-certs\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.588730 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-public-tls-certs\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.589864 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-combined-ca-bundle\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.594057 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-credential-keys\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.596937 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-config-data\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.596960 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-scripts\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.597471 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6tw9\" (UniqueName: \"kubernetes.io/projected/3c94ac11-4a54-4fde-b2dd-e9e92d58894f-kube-api-access-n6tw9\") pod \"keystone-7b7d8889-c8cpc\" (UID: \"3c94ac11-4a54-4fde-b2dd-e9e92d58894f\") " pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.625229 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.636744 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.687959 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kp6v5" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.767944 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.886186 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b2112bdd-0811-4bd0-927d-277b44b5cfc0-config\") pod \"b2112bdd-0811-4bd0-927d-277b44b5cfc0\" (UID: \"b2112bdd-0811-4bd0-927d-277b44b5cfc0\") " Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.886950 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2112bdd-0811-4bd0-927d-277b44b5cfc0-combined-ca-bundle\") pod \"b2112bdd-0811-4bd0-927d-277b44b5cfc0\" (UID: \"b2112bdd-0811-4bd0-927d-277b44b5cfc0\") " Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.886985 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c654c\" (UniqueName: \"kubernetes.io/projected/b2112bdd-0811-4bd0-927d-277b44b5cfc0-kube-api-access-c654c\") pod \"b2112bdd-0811-4bd0-927d-277b44b5cfc0\" (UID: \"b2112bdd-0811-4bd0-927d-277b44b5cfc0\") " Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.891238 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2112bdd-0811-4bd0-927d-277b44b5cfc0-kube-api-access-c654c" (OuterVolumeSpecName: "kube-api-access-c654c") pod "b2112bdd-0811-4bd0-927d-277b44b5cfc0" (UID: "b2112bdd-0811-4bd0-927d-277b44b5cfc0"). InnerVolumeSpecName "kube-api-access-c654c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.920684 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2112bdd-0811-4bd0-927d-277b44b5cfc0-config" (OuterVolumeSpecName: "config") pod "b2112bdd-0811-4bd0-927d-277b44b5cfc0" (UID: "b2112bdd-0811-4bd0-927d-277b44b5cfc0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.922067 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2112bdd-0811-4bd0-927d-277b44b5cfc0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2112bdd-0811-4bd0-927d-277b44b5cfc0" (UID: "b2112bdd-0811-4bd0-927d-277b44b5cfc0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.989766 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2112bdd-0811-4bd0-927d-277b44b5cfc0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.989806 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c654c\" (UniqueName: \"kubernetes.io/projected/b2112bdd-0811-4bd0-927d-277b44b5cfc0-kube-api-access-c654c\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:40 crc kubenswrapper[4808]: I1124 17:43:40.989819 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b2112bdd-0811-4bd0-927d-277b44b5cfc0-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.213906 4808 generic.go:334] "Generic (PLEG): container finished" podID="67ce2c95-e3aa-4d5f-8597-55c293b9e22b" containerID="b91696c1a899298ddd07d20470b57336c1acdd027253fcf40687a2cb526de70c" exitCode=0 Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.213947 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wjrlj" event={"ID":"67ce2c95-e3aa-4d5f-8597-55c293b9e22b","Type":"ContainerDied","Data":"b91696c1a899298ddd07d20470b57336c1acdd027253fcf40687a2cb526de70c"} Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.218238 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kp6v5" event={"ID":"b2112bdd-0811-4bd0-927d-277b44b5cfc0","Type":"ContainerDied","Data":"634201869afaeb2c28c5ff085a585872681ba63052cd5fa1f81b0eb5b54a8000"} Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.218283 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="634201869afaeb2c28c5ff085a585872681ba63052cd5fa1f81b0eb5b54a8000" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.218336 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kp6v5" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.219474 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.219511 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 17:43:41 crc kubenswrapper[4808]: W1124 17:43:41.266958 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c94ac11_4a54_4fde_b2dd_e9e92d58894f.slice/crio-395f1200393beb800e8bdf3c22a7049428523dc8cbfa7488a2b1d35eec03fd88 WatchSource:0}: Error finding container 395f1200393beb800e8bdf3c22a7049428523dc8cbfa7488a2b1d35eec03fd88: Status 404 returned error can't find the container with id 395f1200393beb800e8bdf3c22a7049428523dc8cbfa7488a2b1d35eec03fd88 Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.269336 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7b7d8889-c8cpc"] Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.579897 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hw9dh"] Nov 24 17:43:41 crc kubenswrapper[4808]: E1124 17:43:41.583617 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2112bdd-0811-4bd0-927d-277b44b5cfc0" containerName="neutron-db-sync" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.583661 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2112bdd-0811-4bd0-927d-277b44b5cfc0" containerName="neutron-db-sync" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.583870 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2112bdd-0811-4bd0-927d-277b44b5cfc0" containerName="neutron-db-sync" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.585036 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.602360 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hw9dh"] Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.642143 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-84d884d7c8-csbsl"] Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.647002 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.653797 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-nspql" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.653844 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.654140 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.657518 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.659442 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-84d884d7c8-csbsl"] Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.718695 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.718834 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsm57\" (UniqueName: \"kubernetes.io/projected/e12f8ecd-e886-46d1-b63b-1dccb5d361af-kube-api-access-fsm57\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.718877 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-dns-svc\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.718907 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.718958 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.719000 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-config\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.821678 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsm57\" (UniqueName: \"kubernetes.io/projected/e12f8ecd-e886-46d1-b63b-1dccb5d361af-kube-api-access-fsm57\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.822841 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-dns-svc\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.822950 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.823035 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gnns\" (UniqueName: \"kubernetes.io/projected/954e4c03-8c79-4207-8f16-97d8db1c1b25-kube-api-access-5gnns\") pod \"neutron-84d884d7c8-csbsl\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.823086 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-config\") pod \"neutron-84d884d7c8-csbsl\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.823132 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.823166 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-ovndb-tls-certs\") pod \"neutron-84d884d7c8-csbsl\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.823191 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-config\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.823252 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-httpd-config\") pod \"neutron-84d884d7c8-csbsl\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.823488 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.823578 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-combined-ca-bundle\") pod \"neutron-84d884d7c8-csbsl\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.825631 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-config\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.826350 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.826943 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.827691 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.832036 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-dns-svc\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.854970 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsm57\" (UniqueName: \"kubernetes.io/projected/e12f8ecd-e886-46d1-b63b-1dccb5d361af-kube-api-access-fsm57\") pod \"dnsmasq-dns-55f844cf75-hw9dh\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.926333 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gnns\" (UniqueName: \"kubernetes.io/projected/954e4c03-8c79-4207-8f16-97d8db1c1b25-kube-api-access-5gnns\") pod \"neutron-84d884d7c8-csbsl\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.927758 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-config\") pod \"neutron-84d884d7c8-csbsl\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.928213 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-ovndb-tls-certs\") pod \"neutron-84d884d7c8-csbsl\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.928303 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-httpd-config\") pod \"neutron-84d884d7c8-csbsl\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.928390 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-combined-ca-bundle\") pod \"neutron-84d884d7c8-csbsl\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.934854 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-config\") pod \"neutron-84d884d7c8-csbsl\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.935499 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-combined-ca-bundle\") pod \"neutron-84d884d7c8-csbsl\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.940794 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-httpd-config\") pod \"neutron-84d884d7c8-csbsl\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.941493 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-ovndb-tls-certs\") pod \"neutron-84d884d7c8-csbsl\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.946726 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.947060 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gnns\" (UniqueName: \"kubernetes.io/projected/954e4c03-8c79-4207-8f16-97d8db1c1b25-kube-api-access-5gnns\") pod \"neutron-84d884d7c8-csbsl\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:41 crc kubenswrapper[4808]: I1124 17:43:41.984506 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.247447 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74bc8cd64-b5zpt" event={"ID":"da69b7d2-5a35-4dd7-accb-937fefd0d94f","Type":"ContainerStarted","Data":"e9322e86325f42914f21f1bb36ca087701f02f1c1f709ab3cf501aaab652f98f"} Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.247777 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74bc8cd64-b5zpt" event={"ID":"da69b7d2-5a35-4dd7-accb-937fefd0d94f","Type":"ContainerStarted","Data":"cf560a8ffc4af2db3658fa2946d980a7f4fc1a2402bb6f325a229da64eb2fad6"} Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.260273 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-vnqqn" event={"ID":"71142c75-1dcb-4b06-92bd-f9882aa6a929","Type":"ContainerStarted","Data":"425eda6db936f103a37c96001c727487212eb86f3674f8faeb79054849f1e6ba"} Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.274487 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-74bc8cd64-b5zpt" podStartSLOduration=-9223371998.580307 podStartE2EDuration="38.274469142s" podCreationTimestamp="2025-11-24 17:43:04 +0000 UTC" firstStartedPulling="2025-11-24 17:43:04.878614046 +0000 UTC m=+977.476281848" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:43:42.271307679 +0000 UTC m=+1014.868975481" watchObservedRunningTime="2025-11-24 17:43:42.274469142 +0000 UTC m=+1014.872136954" Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.282764 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7b7d8889-c8cpc" event={"ID":"3c94ac11-4a54-4fde-b2dd-e9e92d58894f","Type":"ContainerStarted","Data":"ac902e1707c30f62a7a904a3fac1526f007f44f44ec2eb6f4164acb97be398fe"} Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.282830 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7b7d8889-c8cpc" event={"ID":"3c94ac11-4a54-4fde-b2dd-e9e92d58894f","Type":"ContainerStarted","Data":"395f1200393beb800e8bdf3c22a7049428523dc8cbfa7488a2b1d35eec03fd88"} Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.282852 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.284776 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-vnqqn" podStartSLOduration=2.93258776 podStartE2EDuration="48.284758532s" podCreationTimestamp="2025-11-24 17:42:54 +0000 UTC" firstStartedPulling="2025-11-24 17:42:56.567237996 +0000 UTC m=+969.164905798" lastFinishedPulling="2025-11-24 17:43:41.919408758 +0000 UTC m=+1014.517076570" observedRunningTime="2025-11-24 17:43:42.284189466 +0000 UTC m=+1014.881857268" watchObservedRunningTime="2025-11-24 17:43:42.284758532 +0000 UTC m=+1014.882426344" Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.503893 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7b7d8889-c8cpc" podStartSLOduration=2.503874009 podStartE2EDuration="2.503874009s" podCreationTimestamp="2025-11-24 17:43:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:43:42.328328844 +0000 UTC m=+1014.925996646" watchObservedRunningTime="2025-11-24 17:43:42.503874009 +0000 UTC m=+1015.101541811" Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.509442 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hw9dh"] Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.530448 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-84d884d7c8-csbsl"] Nov 24 17:43:42 crc kubenswrapper[4808]: W1124 17:43:42.536756 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode12f8ecd_e886_46d1_b63b_1dccb5d361af.slice/crio-1c95576fdb8128470b72fe1b36a0e4b5de4879197fb094d96498e4218babd50c WatchSource:0}: Error finding container 1c95576fdb8128470b72fe1b36a0e4b5de4879197fb094d96498e4218babd50c: Status 404 returned error can't find the container with id 1c95576fdb8128470b72fe1b36a0e4b5de4879197fb094d96498e4218babd50c Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.829718 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wjrlj" Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.951446 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-combined-ca-bundle\") pod \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.951606 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59jmn\" (UniqueName: \"kubernetes.io/projected/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-kube-api-access-59jmn\") pod \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.951677 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-scripts\") pod \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.951730 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-logs\") pod \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.951765 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-config-data\") pod \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\" (UID: \"67ce2c95-e3aa-4d5f-8597-55c293b9e22b\") " Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.952348 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-logs" (OuterVolumeSpecName: "logs") pod "67ce2c95-e3aa-4d5f-8597-55c293b9e22b" (UID: "67ce2c95-e3aa-4d5f-8597-55c293b9e22b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.952620 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.960446 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-kube-api-access-59jmn" (OuterVolumeSpecName: "kube-api-access-59jmn") pod "67ce2c95-e3aa-4d5f-8597-55c293b9e22b" (UID: "67ce2c95-e3aa-4d5f-8597-55c293b9e22b"). InnerVolumeSpecName "kube-api-access-59jmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.960553 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-scripts" (OuterVolumeSpecName: "scripts") pod "67ce2c95-e3aa-4d5f-8597-55c293b9e22b" (UID: "67ce2c95-e3aa-4d5f-8597-55c293b9e22b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.981902 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67ce2c95-e3aa-4d5f-8597-55c293b9e22b" (UID: "67ce2c95-e3aa-4d5f-8597-55c293b9e22b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:42 crc kubenswrapper[4808]: I1124 17:43:42.982254 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-config-data" (OuterVolumeSpecName: "config-data") pod "67ce2c95-e3aa-4d5f-8597-55c293b9e22b" (UID: "67ce2c95-e3aa-4d5f-8597-55c293b9e22b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.053906 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.053940 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59jmn\" (UniqueName: \"kubernetes.io/projected/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-kube-api-access-59jmn\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.053953 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.053960 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ce2c95-e3aa-4d5f-8597-55c293b9e22b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.294750 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84d884d7c8-csbsl" event={"ID":"954e4c03-8c79-4207-8f16-97d8db1c1b25","Type":"ContainerStarted","Data":"2ddc7470557ec15c9c9c5b95f36d9892543df0f2d55e01b6d7e3c43669348694"} Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.296146 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" event={"ID":"e12f8ecd-e886-46d1-b63b-1dccb5d361af","Type":"ContainerStarted","Data":"1c95576fdb8128470b72fe1b36a0e4b5de4879197fb094d96498e4218babd50c"} Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.298672 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.298695 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.299502 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wjrlj" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.300716 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wjrlj" event={"ID":"67ce2c95-e3aa-4d5f-8597-55c293b9e22b","Type":"ContainerDied","Data":"6ae460faf1498698ecc56c3c930433f0cd32173d52be88be4a3b708758aed156"} Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.300764 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ae460faf1498698ecc56c3c930433f0cd32173d52be88be4a3b708758aed156" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.387076 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-785cbcd98d-bqfnp"] Nov 24 17:43:43 crc kubenswrapper[4808]: E1124 17:43:43.387617 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67ce2c95-e3aa-4d5f-8597-55c293b9e22b" containerName="placement-db-sync" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.387644 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="67ce2c95-e3aa-4d5f-8597-55c293b9e22b" containerName="placement-db-sync" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.387864 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="67ce2c95-e3aa-4d5f-8597-55c293b9e22b" containerName="placement-db-sync" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.389093 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.393252 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.393393 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.393427 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-4kjnz" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.393539 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.393575 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.410380 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-785cbcd98d-bqfnp"] Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.462195 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18412693-fc60-4860-a2b2-75e830b495a7-logs\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.462684 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/18412693-fc60-4860-a2b2-75e830b495a7-public-tls-certs\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.462758 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18412693-fc60-4860-a2b2-75e830b495a7-combined-ca-bundle\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.462835 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4l4h\" (UniqueName: \"kubernetes.io/projected/18412693-fc60-4860-a2b2-75e830b495a7-kube-api-access-d4l4h\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.463160 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18412693-fc60-4860-a2b2-75e830b495a7-scripts\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.463217 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18412693-fc60-4860-a2b2-75e830b495a7-config-data\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.463318 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/18412693-fc60-4860-a2b2-75e830b495a7-internal-tls-certs\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.565248 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18412693-fc60-4860-a2b2-75e830b495a7-logs\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.565307 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/18412693-fc60-4860-a2b2-75e830b495a7-public-tls-certs\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.565345 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18412693-fc60-4860-a2b2-75e830b495a7-combined-ca-bundle\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.565393 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4l4h\" (UniqueName: \"kubernetes.io/projected/18412693-fc60-4860-a2b2-75e830b495a7-kube-api-access-d4l4h\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.565470 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18412693-fc60-4860-a2b2-75e830b495a7-scripts\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.565499 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18412693-fc60-4860-a2b2-75e830b495a7-config-data\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.565542 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/18412693-fc60-4860-a2b2-75e830b495a7-internal-tls-certs\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.565734 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18412693-fc60-4860-a2b2-75e830b495a7-logs\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.571888 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18412693-fc60-4860-a2b2-75e830b495a7-config-data\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.572257 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/18412693-fc60-4860-a2b2-75e830b495a7-public-tls-certs\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.572610 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18412693-fc60-4860-a2b2-75e830b495a7-scripts\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.573126 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18412693-fc60-4860-a2b2-75e830b495a7-combined-ca-bundle\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.574488 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/18412693-fc60-4860-a2b2-75e830b495a7-internal-tls-certs\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.594723 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4l4h\" (UniqueName: \"kubernetes.io/projected/18412693-fc60-4860-a2b2-75e830b495a7-kube-api-access-d4l4h\") pod \"placement-785cbcd98d-bqfnp\" (UID: \"18412693-fc60-4860-a2b2-75e830b495a7\") " pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.721233 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.909601 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 17:43:43 crc kubenswrapper[4808]: I1124 17:43:43.917679 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.315242 4808 generic.go:334] "Generic (PLEG): container finished" podID="e12f8ecd-e886-46d1-b63b-1dccb5d361af" containerID="338650c3e52011a453fc9a21020d8b174d7ced9348ece5d36edca02a5885d279" exitCode=0 Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.315444 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" event={"ID":"e12f8ecd-e886-46d1-b63b-1dccb5d361af","Type":"ContainerDied","Data":"338650c3e52011a453fc9a21020d8b174d7ced9348ece5d36edca02a5885d279"} Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.317561 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84d884d7c8-csbsl" event={"ID":"954e4c03-8c79-4207-8f16-97d8db1c1b25","Type":"ContainerStarted","Data":"984617fa04d55decfe4978c04508c6eb44700821ad904eb47e074b0b3e0927a2"} Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.392632 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.392690 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.465744 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ff547b7f9-dfqk9"] Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.467436 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.469102 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.475588 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.487948 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ff547b7f9-dfqk9"] Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.539747 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5ddbdb6df8-pv5zj" podUID="96142f0a-3231-4617-bbd2-9f440c73908e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.584636 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-httpd-config\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.584736 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-public-tls-certs\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.584777 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-config\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.584799 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-combined-ca-bundle\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.584825 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv8jd\" (UniqueName: \"kubernetes.io/projected/76a5234c-3951-4c26-92c2-0ead15585dc5-kube-api-access-tv8jd\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.585107 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-ovndb-tls-certs\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.585255 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-internal-tls-certs\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.695283 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-ovndb-tls-certs\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.695371 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-internal-tls-certs\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.695452 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-httpd-config\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.695498 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-public-tls-certs\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.695532 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-config\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.695552 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-combined-ca-bundle\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.695573 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv8jd\" (UniqueName: \"kubernetes.io/projected/76a5234c-3951-4c26-92c2-0ead15585dc5-kube-api-access-tv8jd\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.704116 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-internal-tls-certs\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.704835 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-public-tls-certs\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.706104 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-config\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.707976 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-combined-ca-bundle\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.709665 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-httpd-config\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.711872 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/76a5234c-3951-4c26-92c2-0ead15585dc5-ovndb-tls-certs\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.738884 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv8jd\" (UniqueName: \"kubernetes.io/projected/76a5234c-3951-4c26-92c2-0ead15585dc5-kube-api-access-tv8jd\") pod \"neutron-ff547b7f9-dfqk9\" (UID: \"76a5234c-3951-4c26-92c2-0ead15585dc5\") " pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:44 crc kubenswrapper[4808]: I1124 17:43:44.792195 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:46 crc kubenswrapper[4808]: I1124 17:43:46.347462 4808 generic.go:334] "Generic (PLEG): container finished" podID="71142c75-1dcb-4b06-92bd-f9882aa6a929" containerID="425eda6db936f103a37c96001c727487212eb86f3674f8faeb79054849f1e6ba" exitCode=0 Nov 24 17:43:46 crc kubenswrapper[4808]: I1124 17:43:46.361552 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-vnqqn" event={"ID":"71142c75-1dcb-4b06-92bd-f9882aa6a929","Type":"ContainerDied","Data":"425eda6db936f103a37c96001c727487212eb86f3674f8faeb79054849f1e6ba"} Nov 24 17:43:49 crc kubenswrapper[4808]: I1124 17:43:49.639731 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-vnqqn" Nov 24 17:43:49 crc kubenswrapper[4808]: I1124 17:43:49.796719 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shhg4\" (UniqueName: \"kubernetes.io/projected/71142c75-1dcb-4b06-92bd-f9882aa6a929-kube-api-access-shhg4\") pod \"71142c75-1dcb-4b06-92bd-f9882aa6a929\" (UID: \"71142c75-1dcb-4b06-92bd-f9882aa6a929\") " Nov 24 17:43:49 crc kubenswrapper[4808]: I1124 17:43:49.797307 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71142c75-1dcb-4b06-92bd-f9882aa6a929-combined-ca-bundle\") pod \"71142c75-1dcb-4b06-92bd-f9882aa6a929\" (UID: \"71142c75-1dcb-4b06-92bd-f9882aa6a929\") " Nov 24 17:43:49 crc kubenswrapper[4808]: I1124 17:43:49.797417 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/71142c75-1dcb-4b06-92bd-f9882aa6a929-db-sync-config-data\") pod \"71142c75-1dcb-4b06-92bd-f9882aa6a929\" (UID: \"71142c75-1dcb-4b06-92bd-f9882aa6a929\") " Nov 24 17:43:49 crc kubenswrapper[4808]: I1124 17:43:49.813639 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71142c75-1dcb-4b06-92bd-f9882aa6a929-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "71142c75-1dcb-4b06-92bd-f9882aa6a929" (UID: "71142c75-1dcb-4b06-92bd-f9882aa6a929"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:49 crc kubenswrapper[4808]: I1124 17:43:49.814411 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71142c75-1dcb-4b06-92bd-f9882aa6a929-kube-api-access-shhg4" (OuterVolumeSpecName: "kube-api-access-shhg4") pod "71142c75-1dcb-4b06-92bd-f9882aa6a929" (UID: "71142c75-1dcb-4b06-92bd-f9882aa6a929"). InnerVolumeSpecName "kube-api-access-shhg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:49 crc kubenswrapper[4808]: I1124 17:43:49.871360 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71142c75-1dcb-4b06-92bd-f9882aa6a929-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71142c75-1dcb-4b06-92bd-f9882aa6a929" (UID: "71142c75-1dcb-4b06-92bd-f9882aa6a929"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:49 crc kubenswrapper[4808]: I1124 17:43:49.898925 4808 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/71142c75-1dcb-4b06-92bd-f9882aa6a929-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:49 crc kubenswrapper[4808]: I1124 17:43:49.898958 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shhg4\" (UniqueName: \"kubernetes.io/projected/71142c75-1dcb-4b06-92bd-f9882aa6a929-kube-api-access-shhg4\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:49 crc kubenswrapper[4808]: I1124 17:43:49.898970 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71142c75-1dcb-4b06-92bd-f9882aa6a929-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:49 crc kubenswrapper[4808]: E1124 17:43:49.899557 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="e98950ba-e946-4422-8ecb-e3e5d85b4438" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.016935 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-785cbcd98d-bqfnp"] Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.118420 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ff547b7f9-dfqk9"] Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.382333 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ff547b7f9-dfqk9" event={"ID":"76a5234c-3951-4c26-92c2-0ead15585dc5","Type":"ContainerStarted","Data":"651db641a17822688d9766ba3aabc2bdbd8c14fbbbad28e57706122ef930e7c5"} Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.383946 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84d884d7c8-csbsl" event={"ID":"954e4c03-8c79-4207-8f16-97d8db1c1b25","Type":"ContainerStarted","Data":"93f4fd08374accf2c4ae3ff8ebe704661d5775240687c949c024d9cd639c1ccf"} Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.384218 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.386524 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-785cbcd98d-bqfnp" event={"ID":"18412693-fc60-4860-a2b2-75e830b495a7","Type":"ContainerStarted","Data":"e37f8c525955fcfee01ac9d516d527e0fbcc3c38915cba80f36961b60d64e632"} Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.386553 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-785cbcd98d-bqfnp" event={"ID":"18412693-fc60-4860-a2b2-75e830b495a7","Type":"ContainerStarted","Data":"936008b7f9b4caab5c2d62d0e97da47e1ae945436030227ab3a60b802b268b0d"} Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.389768 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" event={"ID":"e12f8ecd-e886-46d1-b63b-1dccb5d361af","Type":"ContainerStarted","Data":"fd642cfca7b90772c403af3a4b2f735fa7b425de8b634d580d3e90b857125e7b"} Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.392007 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.397484 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-p77zx" event={"ID":"b80d86dd-fbd8-444a-8159-d5c870eb01be","Type":"ContainerStarted","Data":"45c5433e0229e830366205749ed6a0fd4d2ceccb76c0056cba81f64404a6cd3a"} Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.400028 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-vnqqn" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.400028 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-vnqqn" event={"ID":"71142c75-1dcb-4b06-92bd-f9882aa6a929","Type":"ContainerDied","Data":"ea11f82ecb37819ef4c9484584cb0a4ec99cabc8a732fd458b27894260a93eb0"} Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.400262 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea11f82ecb37819ef4c9484584cb0a4ec99cabc8a732fd458b27894260a93eb0" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.408471 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e98950ba-e946-4422-8ecb-e3e5d85b4438","Type":"ContainerStarted","Data":"ca4bc2904e273646e65a7f70ebae9732af408d750b21db9e45107c3e8adab848"} Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.408812 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e98950ba-e946-4422-8ecb-e3e5d85b4438" containerName="ceilometer-notification-agent" containerID="cri-o://7c4b5968a5c94d367cfd0b20b8d880adb83a2f8780d016045aeda94ab1ed7ce9" gracePeriod=30 Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.408931 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.408983 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e98950ba-e946-4422-8ecb-e3e5d85b4438" containerName="proxy-httpd" containerID="cri-o://ca4bc2904e273646e65a7f70ebae9732af408d750b21db9e45107c3e8adab848" gracePeriod=30 Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.409048 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e98950ba-e946-4422-8ecb-e3e5d85b4438" containerName="sg-core" containerID="cri-o://9abef3dccd5ae469421834c4a60c5afa0540d004107a53962e0092a019e66cab" gracePeriod=30 Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.410964 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-84d884d7c8-csbsl" podStartSLOduration=9.410942703 podStartE2EDuration="9.410942703s" podCreationTimestamp="2025-11-24 17:43:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:43:50.408468981 +0000 UTC m=+1023.006136793" watchObservedRunningTime="2025-11-24 17:43:50.410942703 +0000 UTC m=+1023.008610505" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.438499 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" podStartSLOduration=9.438477707 podStartE2EDuration="9.438477707s" podCreationTimestamp="2025-11-24 17:43:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:43:50.433329687 +0000 UTC m=+1023.030997489" watchObservedRunningTime="2025-11-24 17:43:50.438477707 +0000 UTC m=+1023.036145509" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.451362 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-p77zx" podStartSLOduration=2.92412581 podStartE2EDuration="56.451346703s" podCreationTimestamp="2025-11-24 17:42:54 +0000 UTC" firstStartedPulling="2025-11-24 17:42:56.010450977 +0000 UTC m=+968.608118779" lastFinishedPulling="2025-11-24 17:43:49.53767187 +0000 UTC m=+1022.135339672" observedRunningTime="2025-11-24 17:43:50.448810769 +0000 UTC m=+1023.046478571" watchObservedRunningTime="2025-11-24 17:43:50.451346703 +0000 UTC m=+1023.049014505" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.948483 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5db8598f6c-zs7cb"] Nov 24 17:43:50 crc kubenswrapper[4808]: E1124 17:43:50.949151 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71142c75-1dcb-4b06-92bd-f9882aa6a929" containerName="barbican-db-sync" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.949170 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="71142c75-1dcb-4b06-92bd-f9882aa6a929" containerName="barbican-db-sync" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.949417 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="71142c75-1dcb-4b06-92bd-f9882aa6a929" containerName="barbican-db-sync" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.950296 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.962042 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.962243 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-k4l78" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.962808 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 17:43:50 crc kubenswrapper[4808]: I1124 17:43:50.962795 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5db8598f6c-zs7cb"] Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.004825 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7549f65886-b8qq5"] Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.006722 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.013000 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.036995 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7549f65886-b8qq5"] Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.133034 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltj2s\" (UniqueName: \"kubernetes.io/projected/add8319e-0677-426b-b094-25b79b7a77e4-kube-api-access-ltj2s\") pod \"barbican-keystone-listener-7549f65886-b8qq5\" (UID: \"add8319e-0677-426b-b094-25b79b7a77e4\") " pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.133106 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqknf\" (UniqueName: \"kubernetes.io/projected/c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba-kube-api-access-cqknf\") pod \"barbican-worker-5db8598f6c-zs7cb\" (UID: \"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba\") " pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.133176 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/add8319e-0677-426b-b094-25b79b7a77e4-config-data-custom\") pod \"barbican-keystone-listener-7549f65886-b8qq5\" (UID: \"add8319e-0677-426b-b094-25b79b7a77e4\") " pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.133221 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/add8319e-0677-426b-b094-25b79b7a77e4-logs\") pod \"barbican-keystone-listener-7549f65886-b8qq5\" (UID: \"add8319e-0677-426b-b094-25b79b7a77e4\") " pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.133253 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba-combined-ca-bundle\") pod \"barbican-worker-5db8598f6c-zs7cb\" (UID: \"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba\") " pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.133279 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/add8319e-0677-426b-b094-25b79b7a77e4-config-data\") pod \"barbican-keystone-listener-7549f65886-b8qq5\" (UID: \"add8319e-0677-426b-b094-25b79b7a77e4\") " pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.133306 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba-logs\") pod \"barbican-worker-5db8598f6c-zs7cb\" (UID: \"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba\") " pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.133329 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba-config-data-custom\") pod \"barbican-worker-5db8598f6c-zs7cb\" (UID: \"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba\") " pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.133352 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba-config-data\") pod \"barbican-worker-5db8598f6c-zs7cb\" (UID: \"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba\") " pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.134211 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/add8319e-0677-426b-b094-25b79b7a77e4-combined-ca-bundle\") pod \"barbican-keystone-listener-7549f65886-b8qq5\" (UID: \"add8319e-0677-426b-b094-25b79b7a77e4\") " pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.137172 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hw9dh"] Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.152573 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-c2c2m"] Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.154059 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.193929 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-c2c2m"] Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.237129 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/add8319e-0677-426b-b094-25b79b7a77e4-combined-ca-bundle\") pod \"barbican-keystone-listener-7549f65886-b8qq5\" (UID: \"add8319e-0677-426b-b094-25b79b7a77e4\") " pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.237183 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltj2s\" (UniqueName: \"kubernetes.io/projected/add8319e-0677-426b-b094-25b79b7a77e4-kube-api-access-ltj2s\") pod \"barbican-keystone-listener-7549f65886-b8qq5\" (UID: \"add8319e-0677-426b-b094-25b79b7a77e4\") " pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.237213 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqknf\" (UniqueName: \"kubernetes.io/projected/c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba-kube-api-access-cqknf\") pod \"barbican-worker-5db8598f6c-zs7cb\" (UID: \"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba\") " pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.237270 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/add8319e-0677-426b-b094-25b79b7a77e4-config-data-custom\") pod \"barbican-keystone-listener-7549f65886-b8qq5\" (UID: \"add8319e-0677-426b-b094-25b79b7a77e4\") " pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.237308 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/add8319e-0677-426b-b094-25b79b7a77e4-logs\") pod \"barbican-keystone-listener-7549f65886-b8qq5\" (UID: \"add8319e-0677-426b-b094-25b79b7a77e4\") " pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.237355 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba-combined-ca-bundle\") pod \"barbican-worker-5db8598f6c-zs7cb\" (UID: \"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba\") " pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.237387 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/add8319e-0677-426b-b094-25b79b7a77e4-config-data\") pod \"barbican-keystone-listener-7549f65886-b8qq5\" (UID: \"add8319e-0677-426b-b094-25b79b7a77e4\") " pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.237416 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba-logs\") pod \"barbican-worker-5db8598f6c-zs7cb\" (UID: \"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba\") " pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.237442 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba-config-data-custom\") pod \"barbican-worker-5db8598f6c-zs7cb\" (UID: \"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba\") " pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.237466 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba-config-data\") pod \"barbican-worker-5db8598f6c-zs7cb\" (UID: \"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba\") " pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.238797 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/add8319e-0677-426b-b094-25b79b7a77e4-logs\") pod \"barbican-keystone-listener-7549f65886-b8qq5\" (UID: \"add8319e-0677-426b-b094-25b79b7a77e4\") " pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.243320 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba-logs\") pod \"barbican-worker-5db8598f6c-zs7cb\" (UID: \"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba\") " pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.249634 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba-config-data-custom\") pod \"barbican-worker-5db8598f6c-zs7cb\" (UID: \"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba\") " pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.254114 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/add8319e-0677-426b-b094-25b79b7a77e4-combined-ca-bundle\") pod \"barbican-keystone-listener-7549f65886-b8qq5\" (UID: \"add8319e-0677-426b-b094-25b79b7a77e4\") " pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.254317 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba-config-data\") pod \"barbican-worker-5db8598f6c-zs7cb\" (UID: \"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba\") " pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.254965 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba-combined-ca-bundle\") pod \"barbican-worker-5db8598f6c-zs7cb\" (UID: \"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba\") " pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.256303 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/add8319e-0677-426b-b094-25b79b7a77e4-config-data\") pod \"barbican-keystone-listener-7549f65886-b8qq5\" (UID: \"add8319e-0677-426b-b094-25b79b7a77e4\") " pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.258739 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/add8319e-0677-426b-b094-25b79b7a77e4-config-data-custom\") pod \"barbican-keystone-listener-7549f65886-b8qq5\" (UID: \"add8319e-0677-426b-b094-25b79b7a77e4\") " pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.267361 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqknf\" (UniqueName: \"kubernetes.io/projected/c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba-kube-api-access-cqknf\") pod \"barbican-worker-5db8598f6c-zs7cb\" (UID: \"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba\") " pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.271235 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltj2s\" (UniqueName: \"kubernetes.io/projected/add8319e-0677-426b-b094-25b79b7a77e4-kube-api-access-ltj2s\") pod \"barbican-keystone-listener-7549f65886-b8qq5\" (UID: \"add8319e-0677-426b-b094-25b79b7a77e4\") " pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.279399 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5db8598f6c-zs7cb" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.339891 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-85b59b7bdb-t9l7n"] Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.341862 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.342719 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.342832 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.342868 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-config\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.342898 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-dns-svc\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.343051 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.343092 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc2fm\" (UniqueName: \"kubernetes.io/projected/c856d18a-5ee6-453e-9ef5-2da7c43399e5-kube-api-access-bc2fm\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.355503 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.361333 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.362834 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85b59b7bdb-t9l7n"] Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.446367 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.446440 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc2fm\" (UniqueName: \"kubernetes.io/projected/c856d18a-5ee6-453e-9ef5-2da7c43399e5-kube-api-access-bc2fm\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.446504 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dgj2\" (UniqueName: \"kubernetes.io/projected/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-kube-api-access-4dgj2\") pod \"barbican-api-85b59b7bdb-t9l7n\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.446540 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.446573 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-combined-ca-bundle\") pod \"barbican-api-85b59b7bdb-t9l7n\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.446632 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-config-data-custom\") pod \"barbican-api-85b59b7bdb-t9l7n\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.446673 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-logs\") pod \"barbican-api-85b59b7bdb-t9l7n\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.446713 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.446746 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-config\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.446777 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-dns-svc\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.446805 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-config-data\") pod \"barbican-api-85b59b7bdb-t9l7n\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.447972 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.448283 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.449820 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-config\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.450479 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-dns-svc\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.466454 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.483125 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-785cbcd98d-bqfnp" event={"ID":"18412693-fc60-4860-a2b2-75e830b495a7","Type":"ContainerStarted","Data":"c40cfa88ce198bf33e975bbc258b70c0fc902b7b921118db8838095302867784"} Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.483419 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.483443 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.488505 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ff547b7f9-dfqk9" event={"ID":"76a5234c-3951-4c26-92c2-0ead15585dc5","Type":"ContainerStarted","Data":"7cc4081a539501ab0958848bb20ae708167cd90e641666e3f4e72eeba71bb0e1"} Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.510972 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-785cbcd98d-bqfnp" podStartSLOduration=8.510949075 podStartE2EDuration="8.510949075s" podCreationTimestamp="2025-11-24 17:43:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:43:51.502300053 +0000 UTC m=+1024.099967865" watchObservedRunningTime="2025-11-24 17:43:51.510949075 +0000 UTC m=+1024.108616877" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.519947 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc2fm\" (UniqueName: \"kubernetes.io/projected/c856d18a-5ee6-453e-9ef5-2da7c43399e5-kube-api-access-bc2fm\") pod \"dnsmasq-dns-85ff748b95-c2c2m\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.527689 4808 generic.go:334] "Generic (PLEG): container finished" podID="e98950ba-e946-4422-8ecb-e3e5d85b4438" containerID="ca4bc2904e273646e65a7f70ebae9732af408d750b21db9e45107c3e8adab848" exitCode=0 Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.527732 4808 generic.go:334] "Generic (PLEG): container finished" podID="e98950ba-e946-4422-8ecb-e3e5d85b4438" containerID="9abef3dccd5ae469421834c4a60c5afa0540d004107a53962e0092a019e66cab" exitCode=2 Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.528625 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e98950ba-e946-4422-8ecb-e3e5d85b4438","Type":"ContainerDied","Data":"ca4bc2904e273646e65a7f70ebae9732af408d750b21db9e45107c3e8adab848"} Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.528660 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e98950ba-e946-4422-8ecb-e3e5d85b4438","Type":"ContainerDied","Data":"9abef3dccd5ae469421834c4a60c5afa0540d004107a53962e0092a019e66cab"} Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.548303 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-logs\") pod \"barbican-api-85b59b7bdb-t9l7n\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.548400 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-config-data\") pod \"barbican-api-85b59b7bdb-t9l7n\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.548540 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dgj2\" (UniqueName: \"kubernetes.io/projected/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-kube-api-access-4dgj2\") pod \"barbican-api-85b59b7bdb-t9l7n\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.548571 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-combined-ca-bundle\") pod \"barbican-api-85b59b7bdb-t9l7n\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.548630 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-config-data-custom\") pod \"barbican-api-85b59b7bdb-t9l7n\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.550340 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-logs\") pod \"barbican-api-85b59b7bdb-t9l7n\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.557359 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-config-data\") pod \"barbican-api-85b59b7bdb-t9l7n\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.559293 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-combined-ca-bundle\") pod \"barbican-api-85b59b7bdb-t9l7n\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.561331 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-config-data-custom\") pod \"barbican-api-85b59b7bdb-t9l7n\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.570726 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dgj2\" (UniqueName: \"kubernetes.io/projected/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-kube-api-access-4dgj2\") pod \"barbican-api-85b59b7bdb-t9l7n\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.670524 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:51 crc kubenswrapper[4808]: I1124 17:43:51.802942 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:52 crc kubenswrapper[4808]: I1124 17:43:52.039360 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5db8598f6c-zs7cb"] Nov 24 17:43:52 crc kubenswrapper[4808]: I1124 17:43:52.253245 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7549f65886-b8qq5"] Nov 24 17:43:52 crc kubenswrapper[4808]: I1124 17:43:52.362833 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85b59b7bdb-t9l7n"] Nov 24 17:43:52 crc kubenswrapper[4808]: W1124 17:43:52.363396 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3b80cf2_f210_4ed6_96aa_aa86f42ad126.slice/crio-41c916f5da07218719e52dd8324244b1dc39ef669f9dc11624e624faadb40779 WatchSource:0}: Error finding container 41c916f5da07218719e52dd8324244b1dc39ef669f9dc11624e624faadb40779: Status 404 returned error can't find the container with id 41c916f5da07218719e52dd8324244b1dc39ef669f9dc11624e624faadb40779 Nov 24 17:43:52 crc kubenswrapper[4808]: I1124 17:43:52.449045 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-c2c2m"] Nov 24 17:43:52 crc kubenswrapper[4808]: I1124 17:43:52.560564 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5db8598f6c-zs7cb" event={"ID":"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba","Type":"ContainerStarted","Data":"ec8c9f01593fcbf8e7b46c195f4f2b06727e3bdc9f9404dd04d74c5bc49864b0"} Nov 24 17:43:52 crc kubenswrapper[4808]: I1124 17:43:52.561765 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" event={"ID":"c856d18a-5ee6-453e-9ef5-2da7c43399e5","Type":"ContainerStarted","Data":"b48a38925fd5c8a967128a4879c47762b3a6fd0109e42d42aa856784a45b7982"} Nov 24 17:43:52 crc kubenswrapper[4808]: I1124 17:43:52.563848 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ff547b7f9-dfqk9" event={"ID":"76a5234c-3951-4c26-92c2-0ead15585dc5","Type":"ContainerStarted","Data":"e2c3e3d9dd94b4bb556a5ed69af609d8183b8487e600fc7d450f8e3a422801da"} Nov 24 17:43:52 crc kubenswrapper[4808]: I1124 17:43:52.565184 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85b59b7bdb-t9l7n" event={"ID":"f3b80cf2-f210-4ed6-96aa-aa86f42ad126","Type":"ContainerStarted","Data":"41c916f5da07218719e52dd8324244b1dc39ef669f9dc11624e624faadb40779"} Nov 24 17:43:52 crc kubenswrapper[4808]: I1124 17:43:52.566169 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" event={"ID":"add8319e-0677-426b-b094-25b79b7a77e4","Type":"ContainerStarted","Data":"415d3b06cb1f74714c7baf1f7fa18d6a80116ec05e725475b15e165574d29c44"} Nov 24 17:43:52 crc kubenswrapper[4808]: I1124 17:43:52.566520 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" podUID="e12f8ecd-e886-46d1-b63b-1dccb5d361af" containerName="dnsmasq-dns" containerID="cri-o://fd642cfca7b90772c403af3a4b2f735fa7b425de8b634d580d3e90b857125e7b" gracePeriod=10 Nov 24 17:43:52 crc kubenswrapper[4808]: I1124 17:43:52.593086 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-ff547b7f9-dfqk9" podStartSLOduration=8.592998103 podStartE2EDuration="8.592998103s" podCreationTimestamp="2025-11-24 17:43:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:43:52.587207274 +0000 UTC m=+1025.184875076" watchObservedRunningTime="2025-11-24 17:43:52.592998103 +0000 UTC m=+1025.190665925" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.038513 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.204718 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-ovsdbserver-nb\") pod \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.205024 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-ovsdbserver-sb\") pod \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.205055 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-dns-svc\") pod \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.205148 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-dns-swift-storage-0\") pod \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.205216 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-config\") pod \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.205270 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsm57\" (UniqueName: \"kubernetes.io/projected/e12f8ecd-e886-46d1-b63b-1dccb5d361af-kube-api-access-fsm57\") pod \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\" (UID: \"e12f8ecd-e886-46d1-b63b-1dccb5d361af\") " Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.219692 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e12f8ecd-e886-46d1-b63b-1dccb5d361af-kube-api-access-fsm57" (OuterVolumeSpecName: "kube-api-access-fsm57") pod "e12f8ecd-e886-46d1-b63b-1dccb5d361af" (UID: "e12f8ecd-e886-46d1-b63b-1dccb5d361af"). InnerVolumeSpecName "kube-api-access-fsm57". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.265195 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e12f8ecd-e886-46d1-b63b-1dccb5d361af" (UID: "e12f8ecd-e886-46d1-b63b-1dccb5d361af"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.268274 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e12f8ecd-e886-46d1-b63b-1dccb5d361af" (UID: "e12f8ecd-e886-46d1-b63b-1dccb5d361af"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.280707 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-config" (OuterVolumeSpecName: "config") pod "e12f8ecd-e886-46d1-b63b-1dccb5d361af" (UID: "e12f8ecd-e886-46d1-b63b-1dccb5d361af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.289491 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e12f8ecd-e886-46d1-b63b-1dccb5d361af" (UID: "e12f8ecd-e886-46d1-b63b-1dccb5d361af"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.290612 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e12f8ecd-e886-46d1-b63b-1dccb5d361af" (UID: "e12f8ecd-e886-46d1-b63b-1dccb5d361af"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.307942 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.307979 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.307992 4808 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.308004 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.308035 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsm57\" (UniqueName: \"kubernetes.io/projected/e12f8ecd-e886-46d1-b63b-1dccb5d361af-kube-api-access-fsm57\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.308049 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e12f8ecd-e886-46d1-b63b-1dccb5d361af-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.602743 4808 generic.go:334] "Generic (PLEG): container finished" podID="e12f8ecd-e886-46d1-b63b-1dccb5d361af" containerID="fd642cfca7b90772c403af3a4b2f735fa7b425de8b634d580d3e90b857125e7b" exitCode=0 Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.603124 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" event={"ID":"e12f8ecd-e886-46d1-b63b-1dccb5d361af","Type":"ContainerDied","Data":"fd642cfca7b90772c403af3a4b2f735fa7b425de8b634d580d3e90b857125e7b"} Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.603158 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" event={"ID":"e12f8ecd-e886-46d1-b63b-1dccb5d361af","Type":"ContainerDied","Data":"1c95576fdb8128470b72fe1b36a0e4b5de4879197fb094d96498e4218babd50c"} Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.603177 4808 scope.go:117] "RemoveContainer" containerID="fd642cfca7b90772c403af3a4b2f735fa7b425de8b634d580d3e90b857125e7b" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.603321 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-hw9dh" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.617042 4808 generic.go:334] "Generic (PLEG): container finished" podID="c856d18a-5ee6-453e-9ef5-2da7c43399e5" containerID="d392f76af1dfd199876d75112c0ba2b38141e23473d060a98d76c4d66afa9148" exitCode=0 Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.617122 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" event={"ID":"c856d18a-5ee6-453e-9ef5-2da7c43399e5","Type":"ContainerDied","Data":"d392f76af1dfd199876d75112c0ba2b38141e23473d060a98d76c4d66afa9148"} Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.624631 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85b59b7bdb-t9l7n" event={"ID":"f3b80cf2-f210-4ed6-96aa-aa86f42ad126","Type":"ContainerStarted","Data":"43831f50d8a8e562a079fb72a9a9864256b0f276f0ec7a57c29e2d45ca7f6787"} Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.624700 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85b59b7bdb-t9l7n" event={"ID":"f3b80cf2-f210-4ed6-96aa-aa86f42ad126","Type":"ContainerStarted","Data":"76134779ad2154132c6560d2503ed81cb0443e10b7c5aa1a886deef1d89d21a2"} Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.624715 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.624904 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.624969 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.677377 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-85b59b7bdb-t9l7n" podStartSLOduration=2.677355117 podStartE2EDuration="2.677355117s" podCreationTimestamp="2025-11-24 17:43:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:43:53.670312321 +0000 UTC m=+1026.267980143" watchObservedRunningTime="2025-11-24 17:43:53.677355117 +0000 UTC m=+1026.275022919" Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.698270 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hw9dh"] Nov 24 17:43:53 crc kubenswrapper[4808]: I1124 17:43:53.715046 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hw9dh"] Nov 24 17:43:53 crc kubenswrapper[4808]: E1124 17:43:53.774235 4808 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode12f8ecd_e886_46d1_b63b_1dccb5d361af.slice/crio-1c95576fdb8128470b72fe1b36a0e4b5de4879197fb094d96498e4218babd50c\": RecentStats: unable to find data in memory cache]" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.278990 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-67f98cdffd-vs7fq"] Nov 24 17:43:54 crc kubenswrapper[4808]: E1124 17:43:54.279619 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e12f8ecd-e886-46d1-b63b-1dccb5d361af" containerName="dnsmasq-dns" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.279634 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e12f8ecd-e886-46d1-b63b-1dccb5d361af" containerName="dnsmasq-dns" Nov 24 17:43:54 crc kubenswrapper[4808]: E1124 17:43:54.279649 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e12f8ecd-e886-46d1-b63b-1dccb5d361af" containerName="init" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.279654 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e12f8ecd-e886-46d1-b63b-1dccb5d361af" containerName="init" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.279865 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e12f8ecd-e886-46d1-b63b-1dccb5d361af" containerName="dnsmasq-dns" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.280847 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.283947 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.284223 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.301620 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-67f98cdffd-vs7fq"] Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.363612 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e12f8ecd-e886-46d1-b63b-1dccb5d361af" path="/var/lib/kubelet/pods/e12f8ecd-e886-46d1-b63b-1dccb5d361af/volumes" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.397875 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-74bc8cd64-b5zpt" podUID="da69b7d2-5a35-4dd7-accb-937fefd0d94f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.430767 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-combined-ca-bundle\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.430868 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-logs\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.430931 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-internal-tls-certs\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.430974 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-config-data-custom\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.431007 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-config-data\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.431105 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqx89\" (UniqueName: \"kubernetes.io/projected/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-kube-api-access-mqx89\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.431135 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-public-tls-certs\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.519678 4808 scope.go:117] "RemoveContainer" containerID="338650c3e52011a453fc9a21020d8b174d7ced9348ece5d36edca02a5885d279" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.532991 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-logs\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.533533 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5ddbdb6df8-pv5zj" podUID="96142f0a-3231-4617-bbd2-9f440c73908e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.534183 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-logs\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.534346 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-internal-tls-certs\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.534431 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-config-data-custom\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.536287 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-config-data\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.536475 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqx89\" (UniqueName: \"kubernetes.io/projected/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-kube-api-access-mqx89\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.536536 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-public-tls-certs\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.536686 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-combined-ca-bundle\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.541669 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-internal-tls-certs\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.545433 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-config-data-custom\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.546034 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-combined-ca-bundle\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.548379 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-config-data\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.554714 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-public-tls-certs\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.554909 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqx89\" (UniqueName: \"kubernetes.io/projected/45f1a3c5-b0ad-429f-a2ff-5d19a23ca362-kube-api-access-mqx89\") pod \"barbican-api-67f98cdffd-vs7fq\" (UID: \"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362\") " pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.567850 4808 scope.go:117] "RemoveContainer" containerID="fd642cfca7b90772c403af3a4b2f735fa7b425de8b634d580d3e90b857125e7b" Nov 24 17:43:54 crc kubenswrapper[4808]: E1124 17:43:54.573806 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd642cfca7b90772c403af3a4b2f735fa7b425de8b634d580d3e90b857125e7b\": container with ID starting with fd642cfca7b90772c403af3a4b2f735fa7b425de8b634d580d3e90b857125e7b not found: ID does not exist" containerID="fd642cfca7b90772c403af3a4b2f735fa7b425de8b634d580d3e90b857125e7b" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.573867 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd642cfca7b90772c403af3a4b2f735fa7b425de8b634d580d3e90b857125e7b"} err="failed to get container status \"fd642cfca7b90772c403af3a4b2f735fa7b425de8b634d580d3e90b857125e7b\": rpc error: code = NotFound desc = could not find container \"fd642cfca7b90772c403af3a4b2f735fa7b425de8b634d580d3e90b857125e7b\": container with ID starting with fd642cfca7b90772c403af3a4b2f735fa7b425de8b634d580d3e90b857125e7b not found: ID does not exist" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.573897 4808 scope.go:117] "RemoveContainer" containerID="338650c3e52011a453fc9a21020d8b174d7ced9348ece5d36edca02a5885d279" Nov 24 17:43:54 crc kubenswrapper[4808]: E1124 17:43:54.574331 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"338650c3e52011a453fc9a21020d8b174d7ced9348ece5d36edca02a5885d279\": container with ID starting with 338650c3e52011a453fc9a21020d8b174d7ced9348ece5d36edca02a5885d279 not found: ID does not exist" containerID="338650c3e52011a453fc9a21020d8b174d7ced9348ece5d36edca02a5885d279" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.574376 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"338650c3e52011a453fc9a21020d8b174d7ced9348ece5d36edca02a5885d279"} err="failed to get container status \"338650c3e52011a453fc9a21020d8b174d7ced9348ece5d36edca02a5885d279\": rpc error: code = NotFound desc = could not find container \"338650c3e52011a453fc9a21020d8b174d7ced9348ece5d36edca02a5885d279\": container with ID starting with 338650c3e52011a453fc9a21020d8b174d7ced9348ece5d36edca02a5885d279 not found: ID does not exist" Nov 24 17:43:54 crc kubenswrapper[4808]: I1124 17:43:54.602085 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.137327 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-67f98cdffd-vs7fq"] Nov 24 17:43:55 crc kubenswrapper[4808]: W1124 17:43:55.143712 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45f1a3c5_b0ad_429f_a2ff_5d19a23ca362.slice/crio-863cec679e7e8c8d80bb9688c4759da66a34f78352206bf9a20451e55dbec8cc WatchSource:0}: Error finding container 863cec679e7e8c8d80bb9688c4759da66a34f78352206bf9a20451e55dbec8cc: Status 404 returned error can't find the container with id 863cec679e7e8c8d80bb9688c4759da66a34f78352206bf9a20451e55dbec8cc Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.622296 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.669842 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67f98cdffd-vs7fq" event={"ID":"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362","Type":"ContainerStarted","Data":"f22ef5148eda045e51016719e6dc01a7553b0669cfa4e4889b913fe64f4b3368"} Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.670124 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67f98cdffd-vs7fq" event={"ID":"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362","Type":"ContainerStarted","Data":"863cec679e7e8c8d80bb9688c4759da66a34f78352206bf9a20451e55dbec8cc"} Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.705305 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" event={"ID":"add8319e-0677-426b-b094-25b79b7a77e4","Type":"ContainerStarted","Data":"78e04e9c7b858f263d44abb849f35fd6226a4900dda2c57c6f7bbef7755c80b0"} Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.705682 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" event={"ID":"add8319e-0677-426b-b094-25b79b7a77e4","Type":"ContainerStarted","Data":"ba42d5ac0d7f5254c51dc8db91b1962815f4fcf54012e9cc8fd83d5fc26d0421"} Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.719822 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5db8598f6c-zs7cb" event={"ID":"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba","Type":"ContainerStarted","Data":"beb6f221628e09a23c5cc7f9ccdff2dbab36320aef18511e36cb945b8456738f"} Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.719872 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5db8598f6c-zs7cb" event={"ID":"c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba","Type":"ContainerStarted","Data":"7e7a0b787113277e91414d18b19a6e1e1a1d17e15e013ac29e20a548b58e3335"} Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.725108 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" event={"ID":"c856d18a-5ee6-453e-9ef5-2da7c43399e5","Type":"ContainerStarted","Data":"28d8b18ff7830377f2cc146132f6863afc224f7ee63b89a87d072c02a6397249"} Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.725161 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.739418 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7549f65886-b8qq5" podStartSLOduration=3.425858056 podStartE2EDuration="5.739400063s" podCreationTimestamp="2025-11-24 17:43:50 +0000 UTC" firstStartedPulling="2025-11-24 17:43:52.267311845 +0000 UTC m=+1024.864979647" lastFinishedPulling="2025-11-24 17:43:54.580853852 +0000 UTC m=+1027.178521654" observedRunningTime="2025-11-24 17:43:55.729828714 +0000 UTC m=+1028.327496536" watchObservedRunningTime="2025-11-24 17:43:55.739400063 +0000 UTC m=+1028.337067855" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.749782 4808 generic.go:334] "Generic (PLEG): container finished" podID="e98950ba-e946-4422-8ecb-e3e5d85b4438" containerID="7c4b5968a5c94d367cfd0b20b8d880adb83a2f8780d016045aeda94ab1ed7ce9" exitCode=0 Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.749833 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e98950ba-e946-4422-8ecb-e3e5d85b4438","Type":"ContainerDied","Data":"7c4b5968a5c94d367cfd0b20b8d880adb83a2f8780d016045aeda94ab1ed7ce9"} Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.749861 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e98950ba-e946-4422-8ecb-e3e5d85b4438","Type":"ContainerDied","Data":"07368dcfe91bf73acafeb65b73c9a3e60b7b7f3e9e3ddbced1a27c4b97d753ca"} Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.749879 4808 scope.go:117] "RemoveContainer" containerID="ca4bc2904e273646e65a7f70ebae9732af408d750b21db9e45107c3e8adab848" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.750080 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.756878 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5db8598f6c-zs7cb" podStartSLOduration=3.235609293 podStartE2EDuration="5.756855483s" podCreationTimestamp="2025-11-24 17:43:50 +0000 UTC" firstStartedPulling="2025-11-24 17:43:52.060927231 +0000 UTC m=+1024.658595033" lastFinishedPulling="2025-11-24 17:43:54.582173411 +0000 UTC m=+1027.179841223" observedRunningTime="2025-11-24 17:43:55.745331846 +0000 UTC m=+1028.342999668" watchObservedRunningTime="2025-11-24 17:43:55.756855483 +0000 UTC m=+1028.354523285" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.775398 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" podStartSLOduration=4.775377183 podStartE2EDuration="4.775377183s" podCreationTimestamp="2025-11-24 17:43:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:43:55.764450284 +0000 UTC m=+1028.362118086" watchObservedRunningTime="2025-11-24 17:43:55.775377183 +0000 UTC m=+1028.373044985" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.782781 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-scripts\") pod \"e98950ba-e946-4422-8ecb-e3e5d85b4438\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.782976 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-sg-core-conf-yaml\") pod \"e98950ba-e946-4422-8ecb-e3e5d85b4438\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.783189 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-config-data\") pod \"e98950ba-e946-4422-8ecb-e3e5d85b4438\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.783447 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dht4q\" (UniqueName: \"kubernetes.io/projected/e98950ba-e946-4422-8ecb-e3e5d85b4438-kube-api-access-dht4q\") pod \"e98950ba-e946-4422-8ecb-e3e5d85b4438\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.783553 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-combined-ca-bundle\") pod \"e98950ba-e946-4422-8ecb-e3e5d85b4438\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.783661 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e98950ba-e946-4422-8ecb-e3e5d85b4438-run-httpd\") pod \"e98950ba-e946-4422-8ecb-e3e5d85b4438\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.783872 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e98950ba-e946-4422-8ecb-e3e5d85b4438-log-httpd\") pod \"e98950ba-e946-4422-8ecb-e3e5d85b4438\" (UID: \"e98950ba-e946-4422-8ecb-e3e5d85b4438\") " Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.784240 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e98950ba-e946-4422-8ecb-e3e5d85b4438-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e98950ba-e946-4422-8ecb-e3e5d85b4438" (UID: "e98950ba-e946-4422-8ecb-e3e5d85b4438"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.785728 4808 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e98950ba-e946-4422-8ecb-e3e5d85b4438-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.789659 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-scripts" (OuterVolumeSpecName: "scripts") pod "e98950ba-e946-4422-8ecb-e3e5d85b4438" (UID: "e98950ba-e946-4422-8ecb-e3e5d85b4438"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.790244 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e98950ba-e946-4422-8ecb-e3e5d85b4438-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e98950ba-e946-4422-8ecb-e3e5d85b4438" (UID: "e98950ba-e946-4422-8ecb-e3e5d85b4438"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.790849 4808 scope.go:117] "RemoveContainer" containerID="9abef3dccd5ae469421834c4a60c5afa0540d004107a53962e0092a019e66cab" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.791243 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e98950ba-e946-4422-8ecb-e3e5d85b4438-kube-api-access-dht4q" (OuterVolumeSpecName: "kube-api-access-dht4q") pod "e98950ba-e946-4422-8ecb-e3e5d85b4438" (UID: "e98950ba-e946-4422-8ecb-e3e5d85b4438"). InnerVolumeSpecName "kube-api-access-dht4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.843346 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e98950ba-e946-4422-8ecb-e3e5d85b4438" (UID: "e98950ba-e946-4422-8ecb-e3e5d85b4438"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.845536 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e98950ba-e946-4422-8ecb-e3e5d85b4438" (UID: "e98950ba-e946-4422-8ecb-e3e5d85b4438"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.878565 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-config-data" (OuterVolumeSpecName: "config-data") pod "e98950ba-e946-4422-8ecb-e3e5d85b4438" (UID: "e98950ba-e946-4422-8ecb-e3e5d85b4438"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.887107 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dht4q\" (UniqueName: \"kubernetes.io/projected/e98950ba-e946-4422-8ecb-e3e5d85b4438-kube-api-access-dht4q\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.887345 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.887422 4808 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e98950ba-e946-4422-8ecb-e3e5d85b4438-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.887484 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.887539 4808 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.887597 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e98950ba-e946-4422-8ecb-e3e5d85b4438-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.946975 4808 scope.go:117] "RemoveContainer" containerID="7c4b5968a5c94d367cfd0b20b8d880adb83a2f8780d016045aeda94ab1ed7ce9" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.971807 4808 scope.go:117] "RemoveContainer" containerID="ca4bc2904e273646e65a7f70ebae9732af408d750b21db9e45107c3e8adab848" Nov 24 17:43:55 crc kubenswrapper[4808]: E1124 17:43:55.972345 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca4bc2904e273646e65a7f70ebae9732af408d750b21db9e45107c3e8adab848\": container with ID starting with ca4bc2904e273646e65a7f70ebae9732af408d750b21db9e45107c3e8adab848 not found: ID does not exist" containerID="ca4bc2904e273646e65a7f70ebae9732af408d750b21db9e45107c3e8adab848" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.972431 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca4bc2904e273646e65a7f70ebae9732af408d750b21db9e45107c3e8adab848"} err="failed to get container status \"ca4bc2904e273646e65a7f70ebae9732af408d750b21db9e45107c3e8adab848\": rpc error: code = NotFound desc = could not find container \"ca4bc2904e273646e65a7f70ebae9732af408d750b21db9e45107c3e8adab848\": container with ID starting with ca4bc2904e273646e65a7f70ebae9732af408d750b21db9e45107c3e8adab848 not found: ID does not exist" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.972518 4808 scope.go:117] "RemoveContainer" containerID="9abef3dccd5ae469421834c4a60c5afa0540d004107a53962e0092a019e66cab" Nov 24 17:43:55 crc kubenswrapper[4808]: E1124 17:43:55.972874 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9abef3dccd5ae469421834c4a60c5afa0540d004107a53962e0092a019e66cab\": container with ID starting with 9abef3dccd5ae469421834c4a60c5afa0540d004107a53962e0092a019e66cab not found: ID does not exist" containerID="9abef3dccd5ae469421834c4a60c5afa0540d004107a53962e0092a019e66cab" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.972918 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9abef3dccd5ae469421834c4a60c5afa0540d004107a53962e0092a019e66cab"} err="failed to get container status \"9abef3dccd5ae469421834c4a60c5afa0540d004107a53962e0092a019e66cab\": rpc error: code = NotFound desc = could not find container \"9abef3dccd5ae469421834c4a60c5afa0540d004107a53962e0092a019e66cab\": container with ID starting with 9abef3dccd5ae469421834c4a60c5afa0540d004107a53962e0092a019e66cab not found: ID does not exist" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.972948 4808 scope.go:117] "RemoveContainer" containerID="7c4b5968a5c94d367cfd0b20b8d880adb83a2f8780d016045aeda94ab1ed7ce9" Nov 24 17:43:55 crc kubenswrapper[4808]: E1124 17:43:55.973236 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c4b5968a5c94d367cfd0b20b8d880adb83a2f8780d016045aeda94ab1ed7ce9\": container with ID starting with 7c4b5968a5c94d367cfd0b20b8d880adb83a2f8780d016045aeda94ab1ed7ce9 not found: ID does not exist" containerID="7c4b5968a5c94d367cfd0b20b8d880adb83a2f8780d016045aeda94ab1ed7ce9" Nov 24 17:43:55 crc kubenswrapper[4808]: I1124 17:43:55.973316 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c4b5968a5c94d367cfd0b20b8d880adb83a2f8780d016045aeda94ab1ed7ce9"} err="failed to get container status \"7c4b5968a5c94d367cfd0b20b8d880adb83a2f8780d016045aeda94ab1ed7ce9\": rpc error: code = NotFound desc = could not find container \"7c4b5968a5c94d367cfd0b20b8d880adb83a2f8780d016045aeda94ab1ed7ce9\": container with ID starting with 7c4b5968a5c94d367cfd0b20b8d880adb83a2f8780d016045aeda94ab1ed7ce9 not found: ID does not exist" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.105007 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.127513 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.147394 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:43:56 crc kubenswrapper[4808]: E1124 17:43:56.147774 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98950ba-e946-4422-8ecb-e3e5d85b4438" containerName="sg-core" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.147791 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98950ba-e946-4422-8ecb-e3e5d85b4438" containerName="sg-core" Nov 24 17:43:56 crc kubenswrapper[4808]: E1124 17:43:56.147820 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98950ba-e946-4422-8ecb-e3e5d85b4438" containerName="proxy-httpd" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.147827 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98950ba-e946-4422-8ecb-e3e5d85b4438" containerName="proxy-httpd" Nov 24 17:43:56 crc kubenswrapper[4808]: E1124 17:43:56.147848 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98950ba-e946-4422-8ecb-e3e5d85b4438" containerName="ceilometer-notification-agent" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.147855 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98950ba-e946-4422-8ecb-e3e5d85b4438" containerName="ceilometer-notification-agent" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.148081 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e98950ba-e946-4422-8ecb-e3e5d85b4438" containerName="sg-core" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.148101 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e98950ba-e946-4422-8ecb-e3e5d85b4438" containerName="proxy-httpd" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.148127 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e98950ba-e946-4422-8ecb-e3e5d85b4438" containerName="ceilometer-notification-agent" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.149795 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.154753 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.155166 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.176904 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.294152 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.294231 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.294270 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79bda77a-32e2-4bbf-a304-7e827af401e8-run-httpd\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.294300 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lxcp\" (UniqueName: \"kubernetes.io/projected/79bda77a-32e2-4bbf-a304-7e827af401e8-kube-api-access-7lxcp\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.294346 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-scripts\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.294438 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79bda77a-32e2-4bbf-a304-7e827af401e8-log-httpd\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.294525 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-config-data\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.363334 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e98950ba-e946-4422-8ecb-e3e5d85b4438" path="/var/lib/kubelet/pods/e98950ba-e946-4422-8ecb-e3e5d85b4438/volumes" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.396350 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-config-data\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.396444 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.396493 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.396532 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79bda77a-32e2-4bbf-a304-7e827af401e8-run-httpd\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.396554 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lxcp\" (UniqueName: \"kubernetes.io/projected/79bda77a-32e2-4bbf-a304-7e827af401e8-kube-api-access-7lxcp\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.396592 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-scripts\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.396616 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79bda77a-32e2-4bbf-a304-7e827af401e8-log-httpd\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.397125 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79bda77a-32e2-4bbf-a304-7e827af401e8-log-httpd\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.397700 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79bda77a-32e2-4bbf-a304-7e827af401e8-run-httpd\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.402814 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.403689 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-config-data\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.405650 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-scripts\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.405800 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.416537 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lxcp\" (UniqueName: \"kubernetes.io/projected/79bda77a-32e2-4bbf-a304-7e827af401e8-kube-api-access-7lxcp\") pod \"ceilometer-0\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.472181 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.761865 4808 generic.go:334] "Generic (PLEG): container finished" podID="b80d86dd-fbd8-444a-8159-d5c870eb01be" containerID="45c5433e0229e830366205749ed6a0fd4d2ceccb76c0056cba81f64404a6cd3a" exitCode=0 Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.762059 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-p77zx" event={"ID":"b80d86dd-fbd8-444a-8159-d5c870eb01be","Type":"ContainerDied","Data":"45c5433e0229e830366205749ed6a0fd4d2ceccb76c0056cba81f64404a6cd3a"} Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.766148 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67f98cdffd-vs7fq" event={"ID":"45f1a3c5-b0ad-429f-a2ff-5d19a23ca362","Type":"ContainerStarted","Data":"35166479f24f0daa88621907fe9b8b6bba371280d18fae9d7f5cd2400787e5ad"} Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.766766 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.811341 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-67f98cdffd-vs7fq" podStartSLOduration=2.811297513 podStartE2EDuration="2.811297513s" podCreationTimestamp="2025-11-24 17:43:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:43:56.804067312 +0000 UTC m=+1029.401735134" watchObservedRunningTime="2025-11-24 17:43:56.811297513 +0000 UTC m=+1029.408965305" Nov 24 17:43:56 crc kubenswrapper[4808]: I1124 17:43:56.950778 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:43:57 crc kubenswrapper[4808]: I1124 17:43:57.780804 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79bda77a-32e2-4bbf-a304-7e827af401e8","Type":"ContainerStarted","Data":"9469ceb2a3994298c8cec3b2a29b663e124431ebf84c778b635cda65cdc7bac1"} Nov 24 17:43:57 crc kubenswrapper[4808]: I1124 17:43:57.781150 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79bda77a-32e2-4bbf-a304-7e827af401e8","Type":"ContainerStarted","Data":"3d02da0b803096c40d1d772d68cac23f8867b51a7c0b49ce7936b117bcbbe1a9"} Nov 24 17:43:57 crc kubenswrapper[4808]: I1124 17:43:57.781995 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.218542 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-p77zx" Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.287107 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l74k\" (UniqueName: \"kubernetes.io/projected/b80d86dd-fbd8-444a-8159-d5c870eb01be-kube-api-access-4l74k\") pod \"b80d86dd-fbd8-444a-8159-d5c870eb01be\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.287184 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-config-data\") pod \"b80d86dd-fbd8-444a-8159-d5c870eb01be\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.287318 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-scripts\") pod \"b80d86dd-fbd8-444a-8159-d5c870eb01be\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.287415 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b80d86dd-fbd8-444a-8159-d5c870eb01be-etc-machine-id\") pod \"b80d86dd-fbd8-444a-8159-d5c870eb01be\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.287486 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-db-sync-config-data\") pod \"b80d86dd-fbd8-444a-8159-d5c870eb01be\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.287544 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-combined-ca-bundle\") pod \"b80d86dd-fbd8-444a-8159-d5c870eb01be\" (UID: \"b80d86dd-fbd8-444a-8159-d5c870eb01be\") " Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.287543 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b80d86dd-fbd8-444a-8159-d5c870eb01be-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b80d86dd-fbd8-444a-8159-d5c870eb01be" (UID: "b80d86dd-fbd8-444a-8159-d5c870eb01be"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.287930 4808 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b80d86dd-fbd8-444a-8159-d5c870eb01be-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.292494 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b80d86dd-fbd8-444a-8159-d5c870eb01be-kube-api-access-4l74k" (OuterVolumeSpecName: "kube-api-access-4l74k") pod "b80d86dd-fbd8-444a-8159-d5c870eb01be" (UID: "b80d86dd-fbd8-444a-8159-d5c870eb01be"). InnerVolumeSpecName "kube-api-access-4l74k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.292561 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b80d86dd-fbd8-444a-8159-d5c870eb01be" (UID: "b80d86dd-fbd8-444a-8159-d5c870eb01be"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.308196 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-scripts" (OuterVolumeSpecName: "scripts") pod "b80d86dd-fbd8-444a-8159-d5c870eb01be" (UID: "b80d86dd-fbd8-444a-8159-d5c870eb01be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.324154 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b80d86dd-fbd8-444a-8159-d5c870eb01be" (UID: "b80d86dd-fbd8-444a-8159-d5c870eb01be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.348938 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-config-data" (OuterVolumeSpecName: "config-data") pod "b80d86dd-fbd8-444a-8159-d5c870eb01be" (UID: "b80d86dd-fbd8-444a-8159-d5c870eb01be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.389099 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.389147 4808 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.389160 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.389169 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l74k\" (UniqueName: \"kubernetes.io/projected/b80d86dd-fbd8-444a-8159-d5c870eb01be-kube-api-access-4l74k\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.389179 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80d86dd-fbd8-444a-8159-d5c870eb01be-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.790153 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79bda77a-32e2-4bbf-a304-7e827af401e8","Type":"ContainerStarted","Data":"e23989b3dae6ccfc9f8162a8b675245071adf74a526be0447364a3ca369ae6d4"} Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.792848 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-p77zx" Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.793464 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-p77zx" event={"ID":"b80d86dd-fbd8-444a-8159-d5c870eb01be","Type":"ContainerDied","Data":"1f8cffbcce837fd726f5b5a10a96d244de563377c20124de1a8e71fa836218ba"} Nov 24 17:43:58 crc kubenswrapper[4808]: I1124 17:43:58.793518 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f8cffbcce837fd726f5b5a10a96d244de563377c20124de1a8e71fa836218ba" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.064242 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:43:59 crc kubenswrapper[4808]: E1124 17:43:59.064717 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80d86dd-fbd8-444a-8159-d5c870eb01be" containerName="cinder-db-sync" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.064732 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80d86dd-fbd8-444a-8159-d5c870eb01be" containerName="cinder-db-sync" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.064966 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b80d86dd-fbd8-444a-8159-d5c870eb01be" containerName="cinder-db-sync" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.066170 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.075130 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.075354 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-tjvqn" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.075483 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.075638 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.090241 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.169444 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-c2c2m"] Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.169681 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" podUID="c856d18a-5ee6-453e-9ef5-2da7c43399e5" containerName="dnsmasq-dns" containerID="cri-o://28d8b18ff7830377f2cc146132f6863afc224f7ee63b89a87d072c02a6397249" gracePeriod=10 Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.171248 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.224332 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-config-data\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.224414 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.224447 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.224488 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcvcw\" (UniqueName: \"kubernetes.io/projected/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-kube-api-access-xcvcw\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.224535 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-scripts\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.224573 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.240987 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-j2p9p"] Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.242977 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.267849 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-j2p9p"] Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.326430 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.326466 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-config\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.326505 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.326537 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-config-data\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.326559 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.326583 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.326627 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.326657 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.326689 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gp8l\" (UniqueName: \"kubernetes.io/projected/d02a83ff-86ec-4de5-b9fb-47d5082c387b-kube-api-access-9gp8l\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.326708 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcvcw\" (UniqueName: \"kubernetes.io/projected/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-kube-api-access-xcvcw\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.326732 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-scripts\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.326766 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.327872 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.332615 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.334278 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-scripts\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.336809 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.339160 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-config-data\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.364778 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcvcw\" (UniqueName: \"kubernetes.io/projected/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-kube-api-access-xcvcw\") pod \"cinder-scheduler-0\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.376088 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.377643 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.381843 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.386635 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.428590 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.428660 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-config\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.428716 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.428765 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.428823 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.428973 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gp8l\" (UniqueName: \"kubernetes.io/projected/d02a83ff-86ec-4de5-b9fb-47d5082c387b-kube-api-access-9gp8l\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.430442 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.432677 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.432712 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-config\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.433592 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.432243 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.452281 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gp8l\" (UniqueName: \"kubernetes.io/projected/d02a83ff-86ec-4de5-b9fb-47d5082c387b-kube-api-access-9gp8l\") pod \"dnsmasq-dns-5c9776ccc5-j2p9p\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.498386 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.530210 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6zrg\" (UniqueName: \"kubernetes.io/projected/8245fc9c-90e5-45c9-b469-0abd15edbaaa-kube-api-access-x6zrg\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.530283 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-config-data-custom\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.530348 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8245fc9c-90e5-45c9-b469-0abd15edbaaa-logs\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.530382 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.530422 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-config-data\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.530451 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-scripts\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.530476 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8245fc9c-90e5-45c9-b469-0abd15edbaaa-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.632723 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6zrg\" (UniqueName: \"kubernetes.io/projected/8245fc9c-90e5-45c9-b469-0abd15edbaaa-kube-api-access-x6zrg\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.633135 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-config-data-custom\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.633210 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8245fc9c-90e5-45c9-b469-0abd15edbaaa-logs\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.633258 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.633309 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-config-data\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.633347 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-scripts\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.633376 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8245fc9c-90e5-45c9-b469-0abd15edbaaa-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.637358 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8245fc9c-90e5-45c9-b469-0abd15edbaaa-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.645329 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8245fc9c-90e5-45c9-b469-0abd15edbaaa-logs\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.646590 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.646729 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-scripts\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.649308 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-config-data\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.649768 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-config-data-custom\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.652051 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.655654 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6zrg\" (UniqueName: \"kubernetes.io/projected/8245fc9c-90e5-45c9-b469-0abd15edbaaa-kube-api-access-x6zrg\") pod \"cinder-api-0\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.748941 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.812140 4808 generic.go:334] "Generic (PLEG): container finished" podID="c856d18a-5ee6-453e-9ef5-2da7c43399e5" containerID="28d8b18ff7830377f2cc146132f6863afc224f7ee63b89a87d072c02a6397249" exitCode=0 Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.812379 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" event={"ID":"c856d18a-5ee6-453e-9ef5-2da7c43399e5","Type":"ContainerDied","Data":"28d8b18ff7830377f2cc146132f6863afc224f7ee63b89a87d072c02a6397249"} Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.812438 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" event={"ID":"c856d18a-5ee6-453e-9ef5-2da7c43399e5","Type":"ContainerDied","Data":"b48a38925fd5c8a967128a4879c47762b3a6fd0109e42d42aa856784a45b7982"} Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.812458 4808 scope.go:117] "RemoveContainer" containerID="28d8b18ff7830377f2cc146132f6863afc224f7ee63b89a87d072c02a6397249" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.812463 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-c2c2m" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.817133 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.827255 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79bda77a-32e2-4bbf-a304-7e827af401e8","Type":"ContainerStarted","Data":"099ebd2a2f88f6ffb39327f15dba057a61cec83a54fe57ca5a656e1209ec47a2"} Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.843958 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-config\") pod \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.844015 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-dns-svc\") pod \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.844099 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-ovsdbserver-sb\") pod \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.844210 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bc2fm\" (UniqueName: \"kubernetes.io/projected/c856d18a-5ee6-453e-9ef5-2da7c43399e5-kube-api-access-bc2fm\") pod \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.844256 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-ovsdbserver-nb\") pod \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.844292 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-dns-swift-storage-0\") pod \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\" (UID: \"c856d18a-5ee6-453e-9ef5-2da7c43399e5\") " Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.901426 4808 scope.go:117] "RemoveContainer" containerID="d392f76af1dfd199876d75112c0ba2b38141e23473d060a98d76c4d66afa9148" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.920627 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c856d18a-5ee6-453e-9ef5-2da7c43399e5-kube-api-access-bc2fm" (OuterVolumeSpecName: "kube-api-access-bc2fm") pod "c856d18a-5ee6-453e-9ef5-2da7c43399e5" (UID: "c856d18a-5ee6-453e-9ef5-2da7c43399e5"). InnerVolumeSpecName "kube-api-access-bc2fm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.936998 4808 scope.go:117] "RemoveContainer" containerID="28d8b18ff7830377f2cc146132f6863afc224f7ee63b89a87d072c02a6397249" Nov 24 17:43:59 crc kubenswrapper[4808]: E1124 17:43:59.951242 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28d8b18ff7830377f2cc146132f6863afc224f7ee63b89a87d072c02a6397249\": container with ID starting with 28d8b18ff7830377f2cc146132f6863afc224f7ee63b89a87d072c02a6397249 not found: ID does not exist" containerID="28d8b18ff7830377f2cc146132f6863afc224f7ee63b89a87d072c02a6397249" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.951293 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28d8b18ff7830377f2cc146132f6863afc224f7ee63b89a87d072c02a6397249"} err="failed to get container status \"28d8b18ff7830377f2cc146132f6863afc224f7ee63b89a87d072c02a6397249\": rpc error: code = NotFound desc = could not find container \"28d8b18ff7830377f2cc146132f6863afc224f7ee63b89a87d072c02a6397249\": container with ID starting with 28d8b18ff7830377f2cc146132f6863afc224f7ee63b89a87d072c02a6397249 not found: ID does not exist" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.951318 4808 scope.go:117] "RemoveContainer" containerID="d392f76af1dfd199876d75112c0ba2b38141e23473d060a98d76c4d66afa9148" Nov 24 17:43:59 crc kubenswrapper[4808]: E1124 17:43:59.954251 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d392f76af1dfd199876d75112c0ba2b38141e23473d060a98d76c4d66afa9148\": container with ID starting with d392f76af1dfd199876d75112c0ba2b38141e23473d060a98d76c4d66afa9148 not found: ID does not exist" containerID="d392f76af1dfd199876d75112c0ba2b38141e23473d060a98d76c4d66afa9148" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.954325 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d392f76af1dfd199876d75112c0ba2b38141e23473d060a98d76c4d66afa9148"} err="failed to get container status \"d392f76af1dfd199876d75112c0ba2b38141e23473d060a98d76c4d66afa9148\": rpc error: code = NotFound desc = could not find container \"d392f76af1dfd199876d75112c0ba2b38141e23473d060a98d76c4d66afa9148\": container with ID starting with d392f76af1dfd199876d75112c0ba2b38141e23473d060a98d76c4d66afa9148 not found: ID does not exist" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.957943 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-config" (OuterVolumeSpecName: "config") pod "c856d18a-5ee6-453e-9ef5-2da7c43399e5" (UID: "c856d18a-5ee6-453e-9ef5-2da7c43399e5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.965517 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.965548 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bc2fm\" (UniqueName: \"kubernetes.io/projected/c856d18a-5ee6-453e-9ef5-2da7c43399e5-kube-api-access-bc2fm\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.985364 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c856d18a-5ee6-453e-9ef5-2da7c43399e5" (UID: "c856d18a-5ee6-453e-9ef5-2da7c43399e5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:43:59 crc kubenswrapper[4808]: I1124 17:43:59.988548 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c856d18a-5ee6-453e-9ef5-2da7c43399e5" (UID: "c856d18a-5ee6-453e-9ef5-2da7c43399e5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.001945 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c856d18a-5ee6-453e-9ef5-2da7c43399e5" (UID: "c856d18a-5ee6-453e-9ef5-2da7c43399e5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.002998 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c856d18a-5ee6-453e-9ef5-2da7c43399e5" (UID: "c856d18a-5ee6-453e-9ef5-2da7c43399e5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.067073 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.067126 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.067139 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.067150 4808 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c856d18a-5ee6-453e-9ef5-2da7c43399e5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.218987 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-c2c2m"] Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.251746 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-c2c2m"] Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.301782 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-j2p9p"] Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.318351 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:44:00 crc kubenswrapper[4808]: W1124 17:44:00.336513 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd02a83ff_86ec_4de5_b9fb_47d5082c387b.slice/crio-70a5da4417edd76c05b1382f11b01d3808d02e8cf2c39f404919e98934110a91 WatchSource:0}: Error finding container 70a5da4417edd76c05b1382f11b01d3808d02e8cf2c39f404919e98934110a91: Status 404 returned error can't find the container with id 70a5da4417edd76c05b1382f11b01d3808d02e8cf2c39f404919e98934110a91 Nov 24 17:44:00 crc kubenswrapper[4808]: W1124 17:44:00.339509 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f7df6ad_930f_4cab_a90c_8a4ec319f0ff.slice/crio-3e5b843e06f7b258074fb7b227b4162731581e12bd4d791dfa0cb8b7a1ee857f WatchSource:0}: Error finding container 3e5b843e06f7b258074fb7b227b4162731581e12bd4d791dfa0cb8b7a1ee857f: Status 404 returned error can't find the container with id 3e5b843e06f7b258074fb7b227b4162731581e12bd4d791dfa0cb8b7a1ee857f Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.365842 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c856d18a-5ee6-453e-9ef5-2da7c43399e5" path="/var/lib/kubelet/pods/c856d18a-5ee6-453e-9ef5-2da7c43399e5/volumes" Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.583754 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:44:00 crc kubenswrapper[4808]: W1124 17:44:00.593031 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8245fc9c_90e5_45c9_b469_0abd15edbaaa.slice/crio-3cfd579fc4e1f32802474e9e7bfea8227a130e6d5040d2eaaa5ea514119c0e57 WatchSource:0}: Error finding container 3cfd579fc4e1f32802474e9e7bfea8227a130e6d5040d2eaaa5ea514119c0e57: Status 404 returned error can't find the container with id 3cfd579fc4e1f32802474e9e7bfea8227a130e6d5040d2eaaa5ea514119c0e57 Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.837982 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff","Type":"ContainerStarted","Data":"3e5b843e06f7b258074fb7b227b4162731581e12bd4d791dfa0cb8b7a1ee857f"} Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.840363 4808 generic.go:334] "Generic (PLEG): container finished" podID="d02a83ff-86ec-4de5-b9fb-47d5082c387b" containerID="34652f239e549030e3f689284db06659b83af24b7723483a27e6810f27d80f00" exitCode=0 Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.840415 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" event={"ID":"d02a83ff-86ec-4de5-b9fb-47d5082c387b","Type":"ContainerDied","Data":"34652f239e549030e3f689284db06659b83af24b7723483a27e6810f27d80f00"} Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.840432 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" event={"ID":"d02a83ff-86ec-4de5-b9fb-47d5082c387b","Type":"ContainerStarted","Data":"70a5da4417edd76c05b1382f11b01d3808d02e8cf2c39f404919e98934110a91"} Nov 24 17:44:00 crc kubenswrapper[4808]: I1124 17:44:00.858436 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8245fc9c-90e5-45c9-b469-0abd15edbaaa","Type":"ContainerStarted","Data":"3cfd579fc4e1f32802474e9e7bfea8227a130e6d5040d2eaaa5ea514119c0e57"} Nov 24 17:44:01 crc kubenswrapper[4808]: I1124 17:44:01.431204 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:44:01 crc kubenswrapper[4808]: I1124 17:44:01.878851 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79bda77a-32e2-4bbf-a304-7e827af401e8","Type":"ContainerStarted","Data":"3cd780b72720ae5f11014a2d3ca9c7f99fd586c20e6e6c6cf2717c64ed315d1a"} Nov 24 17:44:01 crc kubenswrapper[4808]: I1124 17:44:01.879228 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 17:44:01 crc kubenswrapper[4808]: I1124 17:44:01.887138 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" event={"ID":"d02a83ff-86ec-4de5-b9fb-47d5082c387b","Type":"ContainerStarted","Data":"ee6fe4ea8c6951db3a9fe5c4b758165bf16ef1a9d8063f6b9028ec870d332b0a"} Nov 24 17:44:01 crc kubenswrapper[4808]: I1124 17:44:01.887962 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:44:01 crc kubenswrapper[4808]: I1124 17:44:01.894835 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8245fc9c-90e5-45c9-b469-0abd15edbaaa","Type":"ContainerStarted","Data":"2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c"} Nov 24 17:44:01 crc kubenswrapper[4808]: I1124 17:44:01.936774 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" podStartSLOduration=2.936758957 podStartE2EDuration="2.936758957s" podCreationTimestamp="2025-11-24 17:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:44:01.935369507 +0000 UTC m=+1034.533037319" watchObservedRunningTime="2025-11-24 17:44:01.936758957 +0000 UTC m=+1034.534426759" Nov 24 17:44:01 crc kubenswrapper[4808]: I1124 17:44:01.949042 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.251853787 podStartE2EDuration="5.948996275s" podCreationTimestamp="2025-11-24 17:43:56 +0000 UTC" firstStartedPulling="2025-11-24 17:43:56.970573313 +0000 UTC m=+1029.568241115" lastFinishedPulling="2025-11-24 17:44:00.667715801 +0000 UTC m=+1033.265383603" observedRunningTime="2025-11-24 17:44:01.907812332 +0000 UTC m=+1034.505480134" watchObservedRunningTime="2025-11-24 17:44:01.948996275 +0000 UTC m=+1034.546664087" Nov 24 17:44:02 crc kubenswrapper[4808]: I1124 17:44:02.962789 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff","Type":"ContainerStarted","Data":"ca89e7fd2557315cfa9022b6bea46b05f7f6dbe7e2c560af4551c3b5e4b7957b"} Nov 24 17:44:02 crc kubenswrapper[4808]: I1124 17:44:02.978808 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8245fc9c-90e5-45c9-b469-0abd15edbaaa","Type":"ContainerStarted","Data":"738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57"} Nov 24 17:44:02 crc kubenswrapper[4808]: I1124 17:44:02.979255 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8245fc9c-90e5-45c9-b469-0abd15edbaaa" containerName="cinder-api-log" containerID="cri-o://2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c" gracePeriod=30 Nov 24 17:44:02 crc kubenswrapper[4808]: I1124 17:44:02.979753 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8245fc9c-90e5-45c9-b469-0abd15edbaaa" containerName="cinder-api" containerID="cri-o://738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57" gracePeriod=30 Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.003163 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.003145728 podStartE2EDuration="4.003145728s" podCreationTimestamp="2025-11-24 17:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:44:03.002985703 +0000 UTC m=+1035.600653515" watchObservedRunningTime="2025-11-24 17:44:03.003145728 +0000 UTC m=+1035.600813530" Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.767346 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.881701 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8245fc9c-90e5-45c9-b469-0abd15edbaaa-etc-machine-id\") pod \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.881833 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8245fc9c-90e5-45c9-b469-0abd15edbaaa-logs\") pod \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.881875 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6zrg\" (UniqueName: \"kubernetes.io/projected/8245fc9c-90e5-45c9-b469-0abd15edbaaa-kube-api-access-x6zrg\") pod \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.881913 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-scripts\") pod \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.881935 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-combined-ca-bundle\") pod \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.881902 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8245fc9c-90e5-45c9-b469-0abd15edbaaa-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8245fc9c-90e5-45c9-b469-0abd15edbaaa" (UID: "8245fc9c-90e5-45c9-b469-0abd15edbaaa"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.882042 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-config-data\") pod \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.882110 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-config-data-custom\") pod \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\" (UID: \"8245fc9c-90e5-45c9-b469-0abd15edbaaa\") " Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.882579 4808 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8245fc9c-90e5-45c9-b469-0abd15edbaaa-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.883268 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8245fc9c-90e5-45c9-b469-0abd15edbaaa-logs" (OuterVolumeSpecName: "logs") pod "8245fc9c-90e5-45c9-b469-0abd15edbaaa" (UID: "8245fc9c-90e5-45c9-b469-0abd15edbaaa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.896390 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8245fc9c-90e5-45c9-b469-0abd15edbaaa-kube-api-access-x6zrg" (OuterVolumeSpecName: "kube-api-access-x6zrg") pod "8245fc9c-90e5-45c9-b469-0abd15edbaaa" (UID: "8245fc9c-90e5-45c9-b469-0abd15edbaaa"). InnerVolumeSpecName "kube-api-access-x6zrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.896892 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-scripts" (OuterVolumeSpecName: "scripts") pod "8245fc9c-90e5-45c9-b469-0abd15edbaaa" (UID: "8245fc9c-90e5-45c9-b469-0abd15edbaaa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.909134 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8245fc9c-90e5-45c9-b469-0abd15edbaaa" (UID: "8245fc9c-90e5-45c9-b469-0abd15edbaaa"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.977164 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8245fc9c-90e5-45c9-b469-0abd15edbaaa" (UID: "8245fc9c-90e5-45c9-b469-0abd15edbaaa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.985389 4808 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.985433 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8245fc9c-90e5-45c9-b469-0abd15edbaaa-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.985448 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6zrg\" (UniqueName: \"kubernetes.io/projected/8245fc9c-90e5-45c9-b469-0abd15edbaaa-kube-api-access-x6zrg\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.985461 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:03 crc kubenswrapper[4808]: I1124 17:44:03.985477 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.016174 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-config-data" (OuterVolumeSpecName: "config-data") pod "8245fc9c-90e5-45c9-b469-0abd15edbaaa" (UID: "8245fc9c-90e5-45c9-b469-0abd15edbaaa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.044285 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff","Type":"ContainerStarted","Data":"568f26773aac3916672cf887d9585d1739321017dc221176838b3d4393034bb7"} Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.069874 4808 generic.go:334] "Generic (PLEG): container finished" podID="8245fc9c-90e5-45c9-b469-0abd15edbaaa" containerID="738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57" exitCode=0 Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.069912 4808 generic.go:334] "Generic (PLEG): container finished" podID="8245fc9c-90e5-45c9-b469-0abd15edbaaa" containerID="2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c" exitCode=143 Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.070206 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.070791 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8245fc9c-90e5-45c9-b469-0abd15edbaaa","Type":"ContainerDied","Data":"738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57"} Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.070824 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8245fc9c-90e5-45c9-b469-0abd15edbaaa","Type":"ContainerDied","Data":"2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c"} Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.070839 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8245fc9c-90e5-45c9-b469-0abd15edbaaa","Type":"ContainerDied","Data":"3cfd579fc4e1f32802474e9e7bfea8227a130e6d5040d2eaaa5ea514119c0e57"} Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.070856 4808 scope.go:117] "RemoveContainer" containerID="738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.079512 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.938446128 podStartE2EDuration="5.079496278s" podCreationTimestamp="2025-11-24 17:43:59 +0000 UTC" firstStartedPulling="2025-11-24 17:44:00.342208098 +0000 UTC m=+1032.939875900" lastFinishedPulling="2025-11-24 17:44:01.483258248 +0000 UTC m=+1034.080926050" observedRunningTime="2025-11-24 17:44:04.060405311 +0000 UTC m=+1036.658073113" watchObservedRunningTime="2025-11-24 17:44:04.079496278 +0000 UTC m=+1036.677164080" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.088528 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8245fc9c-90e5-45c9-b469-0abd15edbaaa-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:04 crc kubenswrapper[4808]: E1124 17:44:04.111642 4808 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8245fc9c_90e5_45c9_b469_0abd15edbaaa.slice\": RecentStats: unable to find data in memory cache]" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.150307 4808 scope.go:117] "RemoveContainer" containerID="2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.165269 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.177351 4808 scope.go:117] "RemoveContainer" containerID="738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57" Nov 24 17:44:04 crc kubenswrapper[4808]: E1124 17:44:04.187238 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57\": container with ID starting with 738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57 not found: ID does not exist" containerID="738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.187306 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57"} err="failed to get container status \"738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57\": rpc error: code = NotFound desc = could not find container \"738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57\": container with ID starting with 738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57 not found: ID does not exist" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.187340 4808 scope.go:117] "RemoveContainer" containerID="2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c" Nov 24 17:44:04 crc kubenswrapper[4808]: E1124 17:44:04.189155 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c\": container with ID starting with 2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c not found: ID does not exist" containerID="2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.189201 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c"} err="failed to get container status \"2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c\": rpc error: code = NotFound desc = could not find container \"2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c\": container with ID starting with 2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c not found: ID does not exist" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.189227 4808 scope.go:117] "RemoveContainer" containerID="738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.194468 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.195703 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57"} err="failed to get container status \"738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57\": rpc error: code = NotFound desc = could not find container \"738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57\": container with ID starting with 738e06f43c3743ffeb6a880a660681ae0974898ebadd7a9cbeece1229bba4b57 not found: ID does not exist" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.195747 4808 scope.go:117] "RemoveContainer" containerID="2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.199073 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c"} err="failed to get container status \"2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c\": rpc error: code = NotFound desc = could not find container \"2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c\": container with ID starting with 2e516988e84fbb8d01fbd4328d8dcab16501797b2830d0c57e87ec441941a90c not found: ID does not exist" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.229984 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:44:04 crc kubenswrapper[4808]: E1124 17:44:04.230609 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c856d18a-5ee6-453e-9ef5-2da7c43399e5" containerName="init" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.230676 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c856d18a-5ee6-453e-9ef5-2da7c43399e5" containerName="init" Nov 24 17:44:04 crc kubenswrapper[4808]: E1124 17:44:04.230762 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c856d18a-5ee6-453e-9ef5-2da7c43399e5" containerName="dnsmasq-dns" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.230820 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c856d18a-5ee6-453e-9ef5-2da7c43399e5" containerName="dnsmasq-dns" Nov 24 17:44:04 crc kubenswrapper[4808]: E1124 17:44:04.230889 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8245fc9c-90e5-45c9-b469-0abd15edbaaa" containerName="cinder-api" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.230960 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8245fc9c-90e5-45c9-b469-0abd15edbaaa" containerName="cinder-api" Nov 24 17:44:04 crc kubenswrapper[4808]: E1124 17:44:04.231056 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8245fc9c-90e5-45c9-b469-0abd15edbaaa" containerName="cinder-api-log" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.231121 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8245fc9c-90e5-45c9-b469-0abd15edbaaa" containerName="cinder-api-log" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.231387 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="8245fc9c-90e5-45c9-b469-0abd15edbaaa" containerName="cinder-api" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.231464 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c856d18a-5ee6-453e-9ef5-2da7c43399e5" containerName="dnsmasq-dns" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.231535 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="8245fc9c-90e5-45c9-b469-0abd15edbaaa" containerName="cinder-api-log" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.232705 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.240554 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.240734 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.240855 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.250302 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.358286 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8245fc9c-90e5-45c9-b469-0abd15edbaaa" path="/var/lib/kubelet/pods/8245fc9c-90e5-45c9-b469-0abd15edbaaa/volumes" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.387479 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.393093 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-74bc8cd64-b5zpt" podUID="da69b7d2-5a35-4dd7-accb-937fefd0d94f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.395092 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-scripts\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.395155 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.395182 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/323bb785-3e02-469e-b169-22109303439e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.395208 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-public-tls-certs\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.395292 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-config-data\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.395334 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj8c7\" (UniqueName: \"kubernetes.io/projected/323bb785-3e02-469e-b169-22109303439e-kube-api-access-bj8c7\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.395384 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.395416 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/323bb785-3e02-469e-b169-22109303439e-logs\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.395459 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-config-data-custom\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.496714 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-config-data-custom\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.496942 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-scripts\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.496976 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.496996 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/323bb785-3e02-469e-b169-22109303439e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.497063 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-public-tls-certs\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.497162 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-config-data\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.497229 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj8c7\" (UniqueName: \"kubernetes.io/projected/323bb785-3e02-469e-b169-22109303439e-kube-api-access-bj8c7\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.497298 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.497350 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/323bb785-3e02-469e-b169-22109303439e-logs\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.497849 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/323bb785-3e02-469e-b169-22109303439e-logs\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.499260 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/323bb785-3e02-469e-b169-22109303439e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.503465 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-public-tls-certs\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.514503 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-scripts\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.514876 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-config-data\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.521601 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-config-data-custom\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.522660 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.527905 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/323bb785-3e02-469e-b169-22109303439e-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.531632 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj8c7\" (UniqueName: \"kubernetes.io/projected/323bb785-3e02-469e-b169-22109303439e-kube-api-access-bj8c7\") pod \"cinder-api-0\" (UID: \"323bb785-3e02-469e-b169-22109303439e\") " pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.561729 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 17:44:04 crc kubenswrapper[4808]: I1124 17:44:04.647142 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 17:44:05 crc kubenswrapper[4808]: I1124 17:44:05.041383 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:44:05 crc kubenswrapper[4808]: I1124 17:44:05.125598 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:44:06 crc kubenswrapper[4808]: I1124 17:44:06.097910 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"323bb785-3e02-469e-b169-22109303439e","Type":"ContainerStarted","Data":"35cac22144e81d5f9a80e2d4acf2cb28ec193050a294758dfbbbed84b2e62167"} Nov 24 17:44:06 crc kubenswrapper[4808]: I1124 17:44:06.098319 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"323bb785-3e02-469e-b169-22109303439e","Type":"ContainerStarted","Data":"6ca8472256e9f1e6d83c31687ba82ebef6d10f3dd14dfea738d264cd097c3323"} Nov 24 17:44:06 crc kubenswrapper[4808]: I1124 17:44:06.431693 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:44:06 crc kubenswrapper[4808]: I1124 17:44:06.522617 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:44:06 crc kubenswrapper[4808]: I1124 17:44:06.523011 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:44:06 crc kubenswrapper[4808]: I1124 17:44:06.523082 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:44:06 crc kubenswrapper[4808]: I1124 17:44:06.523940 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"216625ce434c33017d87034c7bea8e120016af1d6be08176847569ea1821e3e1"} pod="openshift-machine-config-operator/machine-config-daemon-ht87w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:44:06 crc kubenswrapper[4808]: I1124 17:44:06.524005 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" containerID="cri-o://216625ce434c33017d87034c7bea8e120016af1d6be08176847569ea1821e3e1" gracePeriod=600 Nov 24 17:44:06 crc kubenswrapper[4808]: I1124 17:44:06.960142 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-67f98cdffd-vs7fq" Nov 24 17:44:07 crc kubenswrapper[4808]: I1124 17:44:07.037137 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-85b59b7bdb-t9l7n"] Nov 24 17:44:07 crc kubenswrapper[4808]: I1124 17:44:07.037343 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-85b59b7bdb-t9l7n" podUID="f3b80cf2-f210-4ed6-96aa-aa86f42ad126" containerName="barbican-api-log" containerID="cri-o://76134779ad2154132c6560d2503ed81cb0443e10b7c5aa1a886deef1d89d21a2" gracePeriod=30 Nov 24 17:44:07 crc kubenswrapper[4808]: I1124 17:44:07.037759 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-85b59b7bdb-t9l7n" podUID="f3b80cf2-f210-4ed6-96aa-aa86f42ad126" containerName="barbican-api" containerID="cri-o://43831f50d8a8e562a079fb72a9a9864256b0f276f0ec7a57c29e2d45ca7f6787" gracePeriod=30 Nov 24 17:44:07 crc kubenswrapper[4808]: I1124 17:44:07.120194 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"323bb785-3e02-469e-b169-22109303439e","Type":"ContainerStarted","Data":"2b59f930a5159a29c2baf0a9f55211c99bfbfbd6986608d399179cf1ceb5652a"} Nov 24 17:44:07 crc kubenswrapper[4808]: I1124 17:44:07.121510 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 17:44:07 crc kubenswrapper[4808]: I1124 17:44:07.136166 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerID="216625ce434c33017d87034c7bea8e120016af1d6be08176847569ea1821e3e1" exitCode=0 Nov 24 17:44:07 crc kubenswrapper[4808]: I1124 17:44:07.137194 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerDied","Data":"216625ce434c33017d87034c7bea8e120016af1d6be08176847569ea1821e3e1"} Nov 24 17:44:07 crc kubenswrapper[4808]: I1124 17:44:07.137233 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"38d3a2110cebc7bb322ebd71a0ddbf64043c7ffce95b69f693d9dde312b5d138"} Nov 24 17:44:07 crc kubenswrapper[4808]: I1124 17:44:07.137255 4808 scope.go:117] "RemoveContainer" containerID="36f63fea078d49c365c55146d216a74b13586224c5d10cc6c372c7c5a5288461" Nov 24 17:44:07 crc kubenswrapper[4808]: I1124 17:44:07.207116 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.20709117 podStartE2EDuration="3.20709117s" podCreationTimestamp="2025-11-24 17:44:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:44:07.168698769 +0000 UTC m=+1039.766366571" watchObservedRunningTime="2025-11-24 17:44:07.20709117 +0000 UTC m=+1039.804758972" Nov 24 17:44:07 crc kubenswrapper[4808]: I1124 17:44:07.674778 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:44:08 crc kubenswrapper[4808]: I1124 17:44:08.153770 4808 generic.go:334] "Generic (PLEG): container finished" podID="f3b80cf2-f210-4ed6-96aa-aa86f42ad126" containerID="76134779ad2154132c6560d2503ed81cb0443e10b7c5aa1a886deef1d89d21a2" exitCode=143 Nov 24 17:44:08 crc kubenswrapper[4808]: I1124 17:44:08.155098 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85b59b7bdb-t9l7n" event={"ID":"f3b80cf2-f210-4ed6-96aa-aa86f42ad126","Type":"ContainerDied","Data":"76134779ad2154132c6560d2503ed81cb0443e10b7c5aa1a886deef1d89d21a2"} Nov 24 17:44:09 crc kubenswrapper[4808]: I1124 17:44:09.500251 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:44:09 crc kubenswrapper[4808]: I1124 17:44:09.577731 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-4f2w2"] Nov 24 17:44:09 crc kubenswrapper[4808]: I1124 17:44:09.577969 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" podUID="9ee3f387-059b-4ca5-bd55-488b737a0c12" containerName="dnsmasq-dns" containerID="cri-o://9bd29801d8ef589fdf3b93006e8a0418eafd525e3f324cd7540d1c30ee458604" gracePeriod=10 Nov 24 17:44:09 crc kubenswrapper[4808]: I1124 17:44:09.625678 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5ddbdb6df8-pv5zj" Nov 24 17:44:09 crc kubenswrapper[4808]: I1124 17:44:09.702053 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74bc8cd64-b5zpt"] Nov 24 17:44:09 crc kubenswrapper[4808]: I1124 17:44:09.702271 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74bc8cd64-b5zpt" podUID="da69b7d2-5a35-4dd7-accb-937fefd0d94f" containerName="horizon-log" containerID="cri-o://cf560a8ffc4af2db3658fa2946d980a7f4fc1a2402bb6f325a229da64eb2fad6" gracePeriod=30 Nov 24 17:44:09 crc kubenswrapper[4808]: I1124 17:44:09.702673 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74bc8cd64-b5zpt" podUID="da69b7d2-5a35-4dd7-accb-937fefd0d94f" containerName="horizon" containerID="cri-o://e9322e86325f42914f21f1bb36ca087701f02f1c1f709ab3cf501aaab652f98f" gracePeriod=30 Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.007140 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.046872 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.176351 4808 generic.go:334] "Generic (PLEG): container finished" podID="9ee3f387-059b-4ca5-bd55-488b737a0c12" containerID="9bd29801d8ef589fdf3b93006e8a0418eafd525e3f324cd7540d1c30ee458604" exitCode=0 Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.176416 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" event={"ID":"9ee3f387-059b-4ca5-bd55-488b737a0c12","Type":"ContainerDied","Data":"9bd29801d8ef589fdf3b93006e8a0418eafd525e3f324cd7540d1c30ee458604"} Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.176459 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" event={"ID":"9ee3f387-059b-4ca5-bd55-488b737a0c12","Type":"ContainerDied","Data":"2bd382437b8b61dddc6964224b17d1d5af836d0c98787d5e6d1230761623f574"} Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.176471 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bd382437b8b61dddc6964224b17d1d5af836d0c98787d5e6d1230761623f574" Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.178042 4808 generic.go:334] "Generic (PLEG): container finished" podID="da69b7d2-5a35-4dd7-accb-937fefd0d94f" containerID="e9322e86325f42914f21f1bb36ca087701f02f1c1f709ab3cf501aaab652f98f" exitCode=0 Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.178081 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74bc8cd64-b5zpt" event={"ID":"da69b7d2-5a35-4dd7-accb-937fefd0d94f","Type":"ContainerDied","Data":"e9322e86325f42914f21f1bb36ca087701f02f1c1f709ab3cf501aaab652f98f"} Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.178253 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" containerName="cinder-scheduler" containerID="cri-o://ca89e7fd2557315cfa9022b6bea46b05f7f6dbe7e2c560af4551c3b5e4b7957b" gracePeriod=30 Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.178386 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" containerName="probe" containerID="cri-o://568f26773aac3916672cf887d9585d1739321017dc221176838b3d4393034bb7" gracePeriod=30 Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.205170 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.328642 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-ovsdbserver-nb\") pod \"9ee3f387-059b-4ca5-bd55-488b737a0c12\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.328698 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-dns-svc\") pod \"9ee3f387-059b-4ca5-bd55-488b737a0c12\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.328830 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5vjz\" (UniqueName: \"kubernetes.io/projected/9ee3f387-059b-4ca5-bd55-488b737a0c12-kube-api-access-b5vjz\") pod \"9ee3f387-059b-4ca5-bd55-488b737a0c12\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.328908 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-ovsdbserver-sb\") pod \"9ee3f387-059b-4ca5-bd55-488b737a0c12\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.328944 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-dns-swift-storage-0\") pod \"9ee3f387-059b-4ca5-bd55-488b737a0c12\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.328974 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-config\") pod \"9ee3f387-059b-4ca5-bd55-488b737a0c12\" (UID: \"9ee3f387-059b-4ca5-bd55-488b737a0c12\") " Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.335344 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ee3f387-059b-4ca5-bd55-488b737a0c12-kube-api-access-b5vjz" (OuterVolumeSpecName: "kube-api-access-b5vjz") pod "9ee3f387-059b-4ca5-bd55-488b737a0c12" (UID: "9ee3f387-059b-4ca5-bd55-488b737a0c12"). InnerVolumeSpecName "kube-api-access-b5vjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.380192 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-config" (OuterVolumeSpecName: "config") pod "9ee3f387-059b-4ca5-bd55-488b737a0c12" (UID: "9ee3f387-059b-4ca5-bd55-488b737a0c12"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.381423 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9ee3f387-059b-4ca5-bd55-488b737a0c12" (UID: "9ee3f387-059b-4ca5-bd55-488b737a0c12"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.388901 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9ee3f387-059b-4ca5-bd55-488b737a0c12" (UID: "9ee3f387-059b-4ca5-bd55-488b737a0c12"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.399755 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9ee3f387-059b-4ca5-bd55-488b737a0c12" (UID: "9ee3f387-059b-4ca5-bd55-488b737a0c12"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.405122 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9ee3f387-059b-4ca5-bd55-488b737a0c12" (UID: "9ee3f387-059b-4ca5-bd55-488b737a0c12"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.431707 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.431745 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.431759 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5vjz\" (UniqueName: \"kubernetes.io/projected/9ee3f387-059b-4ca5-bd55-488b737a0c12-kube-api-access-b5vjz\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.431791 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.431801 4808 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.431809 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ee3f387-059b-4ca5-bd55-488b737a0c12-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:10 crc kubenswrapper[4808]: I1124 17:44:10.945707 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.043155 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-config-data-custom\") pod \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.043316 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dgj2\" (UniqueName: \"kubernetes.io/projected/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-kube-api-access-4dgj2\") pod \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.043391 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-combined-ca-bundle\") pod \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.043520 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-config-data\") pod \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.043595 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-logs\") pod \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\" (UID: \"f3b80cf2-f210-4ed6-96aa-aa86f42ad126\") " Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.044249 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-logs" (OuterVolumeSpecName: "logs") pod "f3b80cf2-f210-4ed6-96aa-aa86f42ad126" (UID: "f3b80cf2-f210-4ed6-96aa-aa86f42ad126"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.048120 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-kube-api-access-4dgj2" (OuterVolumeSpecName: "kube-api-access-4dgj2") pod "f3b80cf2-f210-4ed6-96aa-aa86f42ad126" (UID: "f3b80cf2-f210-4ed6-96aa-aa86f42ad126"). InnerVolumeSpecName "kube-api-access-4dgj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.054001 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f3b80cf2-f210-4ed6-96aa-aa86f42ad126" (UID: "f3b80cf2-f210-4ed6-96aa-aa86f42ad126"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.083403 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3b80cf2-f210-4ed6-96aa-aa86f42ad126" (UID: "f3b80cf2-f210-4ed6-96aa-aa86f42ad126"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.099987 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-config-data" (OuterVolumeSpecName: "config-data") pod "f3b80cf2-f210-4ed6-96aa-aa86f42ad126" (UID: "f3b80cf2-f210-4ed6-96aa-aa86f42ad126"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.152751 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.152786 4808 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.152799 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dgj2\" (UniqueName: \"kubernetes.io/projected/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-kube-api-access-4dgj2\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.152824 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.152834 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3b80cf2-f210-4ed6-96aa-aa86f42ad126-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.201051 4808 generic.go:334] "Generic (PLEG): container finished" podID="6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" containerID="568f26773aac3916672cf887d9585d1739321017dc221176838b3d4393034bb7" exitCode=0 Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.201137 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff","Type":"ContainerDied","Data":"568f26773aac3916672cf887d9585d1739321017dc221176838b3d4393034bb7"} Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.203773 4808 generic.go:334] "Generic (PLEG): container finished" podID="f3b80cf2-f210-4ed6-96aa-aa86f42ad126" containerID="43831f50d8a8e562a079fb72a9a9864256b0f276f0ec7a57c29e2d45ca7f6787" exitCode=0 Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.203865 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-4f2w2" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.205245 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85b59b7bdb-t9l7n" event={"ID":"f3b80cf2-f210-4ed6-96aa-aa86f42ad126","Type":"ContainerDied","Data":"43831f50d8a8e562a079fb72a9a9864256b0f276f0ec7a57c29e2d45ca7f6787"} Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.205291 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85b59b7bdb-t9l7n" event={"ID":"f3b80cf2-f210-4ed6-96aa-aa86f42ad126","Type":"ContainerDied","Data":"41c916f5da07218719e52dd8324244b1dc39ef669f9dc11624e624faadb40779"} Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.205327 4808 scope.go:117] "RemoveContainer" containerID="43831f50d8a8e562a079fb72a9a9864256b0f276f0ec7a57c29e2d45ca7f6787" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.205262 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85b59b7bdb-t9l7n" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.244880 4808 scope.go:117] "RemoveContainer" containerID="76134779ad2154132c6560d2503ed81cb0443e10b7c5aa1a886deef1d89d21a2" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.258050 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-85b59b7bdb-t9l7n"] Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.265193 4808 scope.go:117] "RemoveContainer" containerID="43831f50d8a8e562a079fb72a9a9864256b0f276f0ec7a57c29e2d45ca7f6787" Nov 24 17:44:11 crc kubenswrapper[4808]: E1124 17:44:11.265698 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43831f50d8a8e562a079fb72a9a9864256b0f276f0ec7a57c29e2d45ca7f6787\": container with ID starting with 43831f50d8a8e562a079fb72a9a9864256b0f276f0ec7a57c29e2d45ca7f6787 not found: ID does not exist" containerID="43831f50d8a8e562a079fb72a9a9864256b0f276f0ec7a57c29e2d45ca7f6787" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.265744 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43831f50d8a8e562a079fb72a9a9864256b0f276f0ec7a57c29e2d45ca7f6787"} err="failed to get container status \"43831f50d8a8e562a079fb72a9a9864256b0f276f0ec7a57c29e2d45ca7f6787\": rpc error: code = NotFound desc = could not find container \"43831f50d8a8e562a079fb72a9a9864256b0f276f0ec7a57c29e2d45ca7f6787\": container with ID starting with 43831f50d8a8e562a079fb72a9a9864256b0f276f0ec7a57c29e2d45ca7f6787 not found: ID does not exist" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.265768 4808 scope.go:117] "RemoveContainer" containerID="76134779ad2154132c6560d2503ed81cb0443e10b7c5aa1a886deef1d89d21a2" Nov 24 17:44:11 crc kubenswrapper[4808]: E1124 17:44:11.266092 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76134779ad2154132c6560d2503ed81cb0443e10b7c5aa1a886deef1d89d21a2\": container with ID starting with 76134779ad2154132c6560d2503ed81cb0443e10b7c5aa1a886deef1d89d21a2 not found: ID does not exist" containerID="76134779ad2154132c6560d2503ed81cb0443e10b7c5aa1a886deef1d89d21a2" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.266149 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76134779ad2154132c6560d2503ed81cb0443e10b7c5aa1a886deef1d89d21a2"} err="failed to get container status \"76134779ad2154132c6560d2503ed81cb0443e10b7c5aa1a886deef1d89d21a2\": rpc error: code = NotFound desc = could not find container \"76134779ad2154132c6560d2503ed81cb0443e10b7c5aa1a886deef1d89d21a2\": container with ID starting with 76134779ad2154132c6560d2503ed81cb0443e10b7c5aa1a886deef1d89d21a2 not found: ID does not exist" Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.274848 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-85b59b7bdb-t9l7n"] Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.283585 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-4f2w2"] Nov 24 17:44:11 crc kubenswrapper[4808]: I1124 17:44:11.291691 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-4f2w2"] Nov 24 17:44:12 crc kubenswrapper[4808]: I1124 17:44:12.055672 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:44:12 crc kubenswrapper[4808]: I1124 17:44:12.358260 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ee3f387-059b-4ca5-bd55-488b737a0c12" path="/var/lib/kubelet/pods/9ee3f387-059b-4ca5-bd55-488b737a0c12/volumes" Nov 24 17:44:12 crc kubenswrapper[4808]: I1124 17:44:12.358887 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3b80cf2-f210-4ed6-96aa-aa86f42ad126" path="/var/lib/kubelet/pods/f3b80cf2-f210-4ed6-96aa-aa86f42ad126/volumes" Nov 24 17:44:12 crc kubenswrapper[4808]: I1124 17:44:12.508139 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7b7d8889-c8cpc" Nov 24 17:44:14 crc kubenswrapper[4808]: E1124 17:44:14.411694 4808 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f7df6ad_930f_4cab_a90c_8a4ec319f0ff.slice/crio-conmon-ca89e7fd2557315cfa9022b6bea46b05f7f6dbe7e2c560af4551c3b5e4b7957b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f7df6ad_930f_4cab_a90c_8a4ec319f0ff.slice/crio-ca89e7fd2557315cfa9022b6bea46b05f7f6dbe7e2c560af4551c3b5e4b7957b.scope\": RecentStats: unable to find data in memory cache]" Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.749385 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.818169 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-config-data-custom\") pod \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.818258 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-etc-machine-id\") pod \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.818296 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-combined-ca-bundle\") pod \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.818322 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-config-data\") pod \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.818375 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-scripts\") pod \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.818416 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" (UID: "6f7df6ad-930f-4cab-a90c-8a4ec319f0ff"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.818541 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcvcw\" (UniqueName: \"kubernetes.io/projected/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-kube-api-access-xcvcw\") pod \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\" (UID: \"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff\") " Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.819673 4808 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.834203 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-scripts" (OuterVolumeSpecName: "scripts") pod "6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" (UID: "6f7df6ad-930f-4cab-a90c-8a4ec319f0ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.834592 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-ff547b7f9-dfqk9" Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.840216 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" (UID: "6f7df6ad-930f-4cab-a90c-8a4ec319f0ff"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.843398 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-kube-api-access-xcvcw" (OuterVolumeSpecName: "kube-api-access-xcvcw") pod "6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" (UID: "6f7df6ad-930f-4cab-a90c-8a4ec319f0ff"). InnerVolumeSpecName "kube-api-access-xcvcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.922315 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.922346 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcvcw\" (UniqueName: \"kubernetes.io/projected/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-kube-api-access-xcvcw\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.922357 4808 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.923136 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-84d884d7c8-csbsl"] Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.923492 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-84d884d7c8-csbsl" podUID="954e4c03-8c79-4207-8f16-97d8db1c1b25" containerName="neutron-api" containerID="cri-o://984617fa04d55decfe4978c04508c6eb44700821ad904eb47e074b0b3e0927a2" gracePeriod=30 Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.925821 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-84d884d7c8-csbsl" podUID="954e4c03-8c79-4207-8f16-97d8db1c1b25" containerName="neutron-httpd" containerID="cri-o://93f4fd08374accf2c4ae3ff8ebe704661d5775240687c949c024d9cd639c1ccf" gracePeriod=30 Nov 24 17:44:14 crc kubenswrapper[4808]: I1124 17:44:14.969083 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" (UID: "6f7df6ad-930f-4cab-a90c-8a4ec319f0ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.024587 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.055143 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-config-data" (OuterVolumeSpecName: "config-data") pod "6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" (UID: "6f7df6ad-930f-4cab-a90c-8a4ec319f0ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.126849 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.139814 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.182579 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-785cbcd98d-bqfnp" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.264948 4808 generic.go:334] "Generic (PLEG): container finished" podID="954e4c03-8c79-4207-8f16-97d8db1c1b25" containerID="93f4fd08374accf2c4ae3ff8ebe704661d5775240687c949c024d9cd639c1ccf" exitCode=0 Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.265092 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84d884d7c8-csbsl" event={"ID":"954e4c03-8c79-4207-8f16-97d8db1c1b25","Type":"ContainerDied","Data":"93f4fd08374accf2c4ae3ff8ebe704661d5775240687c949c024d9cd639c1ccf"} Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.280537 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 17:44:15 crc kubenswrapper[4808]: E1124 17:44:15.280918 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3b80cf2-f210-4ed6-96aa-aa86f42ad126" containerName="barbican-api-log" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.280931 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3b80cf2-f210-4ed6-96aa-aa86f42ad126" containerName="barbican-api-log" Nov 24 17:44:15 crc kubenswrapper[4808]: E1124 17:44:15.280944 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3b80cf2-f210-4ed6-96aa-aa86f42ad126" containerName="barbican-api" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.280950 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3b80cf2-f210-4ed6-96aa-aa86f42ad126" containerName="barbican-api" Nov 24 17:44:15 crc kubenswrapper[4808]: E1124 17:44:15.280959 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" containerName="probe" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.280964 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" containerName="probe" Nov 24 17:44:15 crc kubenswrapper[4808]: E1124 17:44:15.280975 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ee3f387-059b-4ca5-bd55-488b737a0c12" containerName="dnsmasq-dns" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.280980 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ee3f387-059b-4ca5-bd55-488b737a0c12" containerName="dnsmasq-dns" Nov 24 17:44:15 crc kubenswrapper[4808]: E1124 17:44:15.280997 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ee3f387-059b-4ca5-bd55-488b737a0c12" containerName="init" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.281002 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ee3f387-059b-4ca5-bd55-488b737a0c12" containerName="init" Nov 24 17:44:15 crc kubenswrapper[4808]: E1124 17:44:15.281037 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" containerName="cinder-scheduler" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.281045 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" containerName="cinder-scheduler" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.281195 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" containerName="cinder-scheduler" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.281209 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" containerName="probe" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.281220 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3b80cf2-f210-4ed6-96aa-aa86f42ad126" containerName="barbican-api" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.281238 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3b80cf2-f210-4ed6-96aa-aa86f42ad126" containerName="barbican-api-log" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.281249 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ee3f387-059b-4ca5-bd55-488b737a0c12" containerName="dnsmasq-dns" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.281865 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.283139 4808 generic.go:334] "Generic (PLEG): container finished" podID="6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" containerID="ca89e7fd2557315cfa9022b6bea46b05f7f6dbe7e2c560af4551c3b5e4b7957b" exitCode=0 Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.304626 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.304898 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.308525 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.311197 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff","Type":"ContainerDied","Data":"ca89e7fd2557315cfa9022b6bea46b05f7f6dbe7e2c560af4551c3b5e4b7957b"} Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.311298 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6f7df6ad-930f-4cab-a90c-8a4ec319f0ff","Type":"ContainerDied","Data":"3e5b843e06f7b258074fb7b227b4162731581e12bd4d791dfa0cb8b7a1ee857f"} Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.311343 4808 scope.go:117] "RemoveContainer" containerID="568f26773aac3916672cf887d9585d1739321017dc221176838b3d4393034bb7" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.313072 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.318231 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-97g8f" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.334692 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-openstack-config\") pod \"openstackclient\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.334770 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qlm4\" (UniqueName: \"kubernetes.io/projected/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-kube-api-access-9qlm4\") pod \"openstackclient\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.334816 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.334853 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-openstack-config-secret\") pod \"openstackclient\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.401303 4808 scope.go:117] "RemoveContainer" containerID="ca89e7fd2557315cfa9022b6bea46b05f7f6dbe7e2c560af4551c3b5e4b7957b" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.421091 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.436510 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-openstack-config\") pod \"openstackclient\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.436603 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qlm4\" (UniqueName: \"kubernetes.io/projected/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-kube-api-access-9qlm4\") pod \"openstackclient\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.436672 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.436744 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-openstack-config-secret\") pod \"openstackclient\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.438114 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.439508 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-openstack-config\") pod \"openstackclient\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.441252 4808 scope.go:117] "RemoveContainer" containerID="568f26773aac3916672cf887d9585d1739321017dc221176838b3d4393034bb7" Nov 24 17:44:15 crc kubenswrapper[4808]: E1124 17:44:15.442735 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"568f26773aac3916672cf887d9585d1739321017dc221176838b3d4393034bb7\": container with ID starting with 568f26773aac3916672cf887d9585d1739321017dc221176838b3d4393034bb7 not found: ID does not exist" containerID="568f26773aac3916672cf887d9585d1739321017dc221176838b3d4393034bb7" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.442773 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"568f26773aac3916672cf887d9585d1739321017dc221176838b3d4393034bb7"} err="failed to get container status \"568f26773aac3916672cf887d9585d1739321017dc221176838b3d4393034bb7\": rpc error: code = NotFound desc = could not find container \"568f26773aac3916672cf887d9585d1739321017dc221176838b3d4393034bb7\": container with ID starting with 568f26773aac3916672cf887d9585d1739321017dc221176838b3d4393034bb7 not found: ID does not exist" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.442798 4808 scope.go:117] "RemoveContainer" containerID="ca89e7fd2557315cfa9022b6bea46b05f7f6dbe7e2c560af4551c3b5e4b7957b" Nov 24 17:44:15 crc kubenswrapper[4808]: E1124 17:44:15.444056 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca89e7fd2557315cfa9022b6bea46b05f7f6dbe7e2c560af4551c3b5e4b7957b\": container with ID starting with ca89e7fd2557315cfa9022b6bea46b05f7f6dbe7e2c560af4551c3b5e4b7957b not found: ID does not exist" containerID="ca89e7fd2557315cfa9022b6bea46b05f7f6dbe7e2c560af4551c3b5e4b7957b" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.444086 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca89e7fd2557315cfa9022b6bea46b05f7f6dbe7e2c560af4551c3b5e4b7957b"} err="failed to get container status \"ca89e7fd2557315cfa9022b6bea46b05f7f6dbe7e2c560af4551c3b5e4b7957b\": rpc error: code = NotFound desc = could not find container \"ca89e7fd2557315cfa9022b6bea46b05f7f6dbe7e2c560af4551c3b5e4b7957b\": container with ID starting with ca89e7fd2557315cfa9022b6bea46b05f7f6dbe7e2c560af4551c3b5e4b7957b not found: ID does not exist" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.454509 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.455593 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-openstack-config-secret\") pod \"openstackclient\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.456097 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.457804 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.460148 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.471592 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.475633 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qlm4\" (UniqueName: \"kubernetes.io/projected/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-kube-api-access-9qlm4\") pod \"openstackclient\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.539323 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.539603 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.539762 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt4cs\" (UniqueName: \"kubernetes.io/projected/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-kube-api-access-rt4cs\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.539896 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.540035 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-scripts\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.540205 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-config-data\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.613736 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.614382 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.623244 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.641779 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.642036 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt4cs\" (UniqueName: \"kubernetes.io/projected/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-kube-api-access-rt4cs\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.642145 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.642251 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-scripts\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.642358 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-config-data\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.642495 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.642393 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.653708 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-scripts\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.655116 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-config-data\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.657962 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.659469 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.682504 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt4cs\" (UniqueName: \"kubernetes.io/projected/7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a-kube-api-access-rt4cs\") pod \"cinder-scheduler-0\" (UID: \"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a\") " pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.692535 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.694001 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.726670 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.744127 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28hcb\" (UniqueName: \"kubernetes.io/projected/2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c-kube-api-access-28hcb\") pod \"openstackclient\" (UID: \"2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.744258 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c-openstack-config\") pod \"openstackclient\" (UID: \"2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.744322 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c-openstack-config-secret\") pod \"openstackclient\" (UID: \"2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.744358 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.846306 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c-openstack-config-secret\") pod \"openstackclient\" (UID: \"2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.846361 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.846443 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28hcb\" (UniqueName: \"kubernetes.io/projected/2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c-kube-api-access-28hcb\") pod \"openstackclient\" (UID: \"2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.846509 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c-openstack-config\") pod \"openstackclient\" (UID: \"2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.848499 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c-openstack-config\") pod \"openstackclient\" (UID: \"2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.851391 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c-openstack-config-secret\") pod \"openstackclient\" (UID: \"2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.854071 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.861969 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 17:44:15 crc kubenswrapper[4808]: I1124 17:44:15.883191 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28hcb\" (UniqueName: \"kubernetes.io/projected/2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c-kube-api-access-28hcb\") pod \"openstackclient\" (UID: \"2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c\") " pod="openstack/openstackclient" Nov 24 17:44:15 crc kubenswrapper[4808]: E1124 17:44:15.907364 4808 log.go:32] "RunPodSandbox from runtime service failed" err=< Nov 24 17:44:15 crc kubenswrapper[4808]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_0ac047a9-8170-4e45-9d8b-d1cc7bf735cc_0(889a5eb616836886a09d3da1b92958198f7423c7c678b873a60df26606d4cfa0): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"889a5eb616836886a09d3da1b92958198f7423c7c678b873a60df26606d4cfa0" Netns:"/var/run/netns/4a9c5c0a-f836-492c-ad85-f3f3b5d9af2e" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=889a5eb616836886a09d3da1b92958198f7423c7c678b873a60df26606d4cfa0;K8S_POD_UID=0ac047a9-8170-4e45-9d8b-d1cc7bf735cc" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc]: expected pod UID "0ac047a9-8170-4e45-9d8b-d1cc7bf735cc" but got "2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c" from Kube API Nov 24 17:44:15 crc kubenswrapper[4808]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 24 17:44:15 crc kubenswrapper[4808]: > Nov 24 17:44:15 crc kubenswrapper[4808]: E1124 17:44:15.907440 4808 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Nov 24 17:44:15 crc kubenswrapper[4808]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_0ac047a9-8170-4e45-9d8b-d1cc7bf735cc_0(889a5eb616836886a09d3da1b92958198f7423c7c678b873a60df26606d4cfa0): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"889a5eb616836886a09d3da1b92958198f7423c7c678b873a60df26606d4cfa0" Netns:"/var/run/netns/4a9c5c0a-f836-492c-ad85-f3f3b5d9af2e" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=889a5eb616836886a09d3da1b92958198f7423c7c678b873a60df26606d4cfa0;K8S_POD_UID=0ac047a9-8170-4e45-9d8b-d1cc7bf735cc" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc]: expected pod UID "0ac047a9-8170-4e45-9d8b-d1cc7bf735cc" but got "2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c" from Kube API Nov 24 17:44:15 crc kubenswrapper[4808]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 24 17:44:15 crc kubenswrapper[4808]: > pod="openstack/openstackclient" Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.086072 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.307241 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.311515 4808 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="0ac047a9-8170-4e45-9d8b-d1cc7bf735cc" podUID="2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c" Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.374252 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.384034 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f7df6ad-930f-4cab-a90c-8a4ec319f0ff" path="/var/lib/kubelet/pods/6f7df6ad-930f-4cab-a90c-8a4ec319f0ff/volumes" Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.384765 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.461270 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-combined-ca-bundle\") pod \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.461380 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-openstack-config-secret\") pod \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.461423 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-openstack-config\") pod \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.461671 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qlm4\" (UniqueName: \"kubernetes.io/projected/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-kube-api-access-9qlm4\") pod \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\" (UID: \"0ac047a9-8170-4e45-9d8b-d1cc7bf735cc\") " Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.462087 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "0ac047a9-8170-4e45-9d8b-d1cc7bf735cc" (UID: "0ac047a9-8170-4e45-9d8b-d1cc7bf735cc"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.465821 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-kube-api-access-9qlm4" (OuterVolumeSpecName: "kube-api-access-9qlm4") pod "0ac047a9-8170-4e45-9d8b-d1cc7bf735cc" (UID: "0ac047a9-8170-4e45-9d8b-d1cc7bf735cc"). InnerVolumeSpecName "kube-api-access-9qlm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.467822 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ac047a9-8170-4e45-9d8b-d1cc7bf735cc" (UID: "0ac047a9-8170-4e45-9d8b-d1cc7bf735cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.470324 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "0ac047a9-8170-4e45-9d8b-d1cc7bf735cc" (UID: "0ac047a9-8170-4e45-9d8b-d1cc7bf735cc"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.564862 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qlm4\" (UniqueName: \"kubernetes.io/projected/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-kube-api-access-9qlm4\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.565119 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.565130 4808 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.565139 4808 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:16 crc kubenswrapper[4808]: I1124 17:44:16.634982 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 17:44:17 crc kubenswrapper[4808]: I1124 17:44:17.306972 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 17:44:17 crc kubenswrapper[4808]: I1124 17:44:17.403286 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a","Type":"ContainerStarted","Data":"006735b6374ae5dc0a70d80e54959f7cd981d1adfc25f8c2e2ec9acbcfb9e702"} Nov 24 17:44:17 crc kubenswrapper[4808]: I1124 17:44:17.403331 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a","Type":"ContainerStarted","Data":"37ab1d50dbc36eb7f6c99841827862e6e180ea7483d1ece2dba2b0f3574c1894"} Nov 24 17:44:17 crc kubenswrapper[4808]: I1124 17:44:17.413399 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c","Type":"ContainerStarted","Data":"aebdd9825fa211fe922547e1d4478100964c3dbb7e94a42bbc489794b6d6fcff"} Nov 24 17:44:17 crc kubenswrapper[4808]: I1124 17:44:17.433504 4808 generic.go:334] "Generic (PLEG): container finished" podID="954e4c03-8c79-4207-8f16-97d8db1c1b25" containerID="984617fa04d55decfe4978c04508c6eb44700821ad904eb47e074b0b3e0927a2" exitCode=0 Nov 24 17:44:17 crc kubenswrapper[4808]: I1124 17:44:17.433577 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 17:44:17 crc kubenswrapper[4808]: I1124 17:44:17.434390 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84d884d7c8-csbsl" event={"ID":"954e4c03-8c79-4207-8f16-97d8db1c1b25","Type":"ContainerDied","Data":"984617fa04d55decfe4978c04508c6eb44700821ad904eb47e074b0b3e0927a2"} Nov 24 17:44:17 crc kubenswrapper[4808]: I1124 17:44:17.465679 4808 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="0ac047a9-8170-4e45-9d8b-d1cc7bf735cc" podUID="2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c" Nov 24 17:44:17 crc kubenswrapper[4808]: I1124 17:44:17.953657 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.023803 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-ovndb-tls-certs\") pod \"954e4c03-8c79-4207-8f16-97d8db1c1b25\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.024070 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-combined-ca-bundle\") pod \"954e4c03-8c79-4207-8f16-97d8db1c1b25\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.024167 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-httpd-config\") pod \"954e4c03-8c79-4207-8f16-97d8db1c1b25\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.024328 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gnns\" (UniqueName: \"kubernetes.io/projected/954e4c03-8c79-4207-8f16-97d8db1c1b25-kube-api-access-5gnns\") pod \"954e4c03-8c79-4207-8f16-97d8db1c1b25\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.024459 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-config\") pod \"954e4c03-8c79-4207-8f16-97d8db1c1b25\" (UID: \"954e4c03-8c79-4207-8f16-97d8db1c1b25\") " Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.043281 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/954e4c03-8c79-4207-8f16-97d8db1c1b25-kube-api-access-5gnns" (OuterVolumeSpecName: "kube-api-access-5gnns") pod "954e4c03-8c79-4207-8f16-97d8db1c1b25" (UID: "954e4c03-8c79-4207-8f16-97d8db1c1b25"). InnerVolumeSpecName "kube-api-access-5gnns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.049367 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "954e4c03-8c79-4207-8f16-97d8db1c1b25" (UID: "954e4c03-8c79-4207-8f16-97d8db1c1b25"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.086163 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-config" (OuterVolumeSpecName: "config") pod "954e4c03-8c79-4207-8f16-97d8db1c1b25" (UID: "954e4c03-8c79-4207-8f16-97d8db1c1b25"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.120626 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "954e4c03-8c79-4207-8f16-97d8db1c1b25" (UID: "954e4c03-8c79-4207-8f16-97d8db1c1b25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.127845 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gnns\" (UniqueName: \"kubernetes.io/projected/954e4c03-8c79-4207-8f16-97d8db1c1b25-kube-api-access-5gnns\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.127877 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.127890 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.127901 4808 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.189340 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "954e4c03-8c79-4207-8f16-97d8db1c1b25" (UID: "954e4c03-8c79-4207-8f16-97d8db1c1b25"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.229754 4808 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/954e4c03-8c79-4207-8f16-97d8db1c1b25-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.402618 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ac047a9-8170-4e45-9d8b-d1cc7bf735cc" path="/var/lib/kubelet/pods/0ac047a9-8170-4e45-9d8b-d1cc7bf735cc/volumes" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.464697 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a","Type":"ContainerStarted","Data":"362944b531b8a75ee5839c6a99731abd81a221f932d43df8927073db74a61ad6"} Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.475612 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84d884d7c8-csbsl" event={"ID":"954e4c03-8c79-4207-8f16-97d8db1c1b25","Type":"ContainerDied","Data":"2ddc7470557ec15c9c9c5b95f36d9892543df0f2d55e01b6d7e3c43669348694"} Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.475680 4808 scope.go:117] "RemoveContainer" containerID="93f4fd08374accf2c4ae3ff8ebe704661d5775240687c949c024d9cd639c1ccf" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.475885 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84d884d7c8-csbsl" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.488310 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.488292681 podStartE2EDuration="3.488292681s" podCreationTimestamp="2025-11-24 17:44:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:44:18.486735305 +0000 UTC m=+1051.084403117" watchObservedRunningTime="2025-11-24 17:44:18.488292681 +0000 UTC m=+1051.085960493" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.517977 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-84d884d7c8-csbsl"] Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.531634 4808 scope.go:117] "RemoveContainer" containerID="984617fa04d55decfe4978c04508c6eb44700821ad904eb47e074b0b3e0927a2" Nov 24 17:44:18 crc kubenswrapper[4808]: I1124 17:44:18.531857 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-84d884d7c8-csbsl"] Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.815603 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-84b96bd867-hvzp6"] Nov 24 17:44:19 crc kubenswrapper[4808]: E1124 17:44:19.816338 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954e4c03-8c79-4207-8f16-97d8db1c1b25" containerName="neutron-api" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.816355 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="954e4c03-8c79-4207-8f16-97d8db1c1b25" containerName="neutron-api" Nov 24 17:44:19 crc kubenswrapper[4808]: E1124 17:44:19.816371 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954e4c03-8c79-4207-8f16-97d8db1c1b25" containerName="neutron-httpd" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.816381 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="954e4c03-8c79-4207-8f16-97d8db1c1b25" containerName="neutron-httpd" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.816644 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="954e4c03-8c79-4207-8f16-97d8db1c1b25" containerName="neutron-api" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.816668 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="954e4c03-8c79-4207-8f16-97d8db1c1b25" containerName="neutron-httpd" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.817973 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.820642 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.820869 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.821150 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.832979 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-84b96bd867-hvzp6"] Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.967382 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a525134b-d25e-4a38-82ec-c7256c6d6697-public-tls-certs\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.967463 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a525134b-d25e-4a38-82ec-c7256c6d6697-config-data\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.967506 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zrhv\" (UniqueName: \"kubernetes.io/projected/a525134b-d25e-4a38-82ec-c7256c6d6697-kube-api-access-9zrhv\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.967540 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a525134b-d25e-4a38-82ec-c7256c6d6697-etc-swift\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.967598 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a525134b-d25e-4a38-82ec-c7256c6d6697-log-httpd\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.967623 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a525134b-d25e-4a38-82ec-c7256c6d6697-run-httpd\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.967660 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a525134b-d25e-4a38-82ec-c7256c6d6697-combined-ca-bundle\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:19 crc kubenswrapper[4808]: I1124 17:44:19.967734 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a525134b-d25e-4a38-82ec-c7256c6d6697-internal-tls-certs\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.069366 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a525134b-d25e-4a38-82ec-c7256c6d6697-public-tls-certs\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.069452 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a525134b-d25e-4a38-82ec-c7256c6d6697-config-data\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.069489 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zrhv\" (UniqueName: \"kubernetes.io/projected/a525134b-d25e-4a38-82ec-c7256c6d6697-kube-api-access-9zrhv\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.069527 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a525134b-d25e-4a38-82ec-c7256c6d6697-etc-swift\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.069589 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a525134b-d25e-4a38-82ec-c7256c6d6697-log-httpd\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.069618 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a525134b-d25e-4a38-82ec-c7256c6d6697-run-httpd\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.069662 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a525134b-d25e-4a38-82ec-c7256c6d6697-combined-ca-bundle\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.069686 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a525134b-d25e-4a38-82ec-c7256c6d6697-internal-tls-certs\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.071268 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a525134b-d25e-4a38-82ec-c7256c6d6697-run-httpd\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.071556 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a525134b-d25e-4a38-82ec-c7256c6d6697-log-httpd\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.076330 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a525134b-d25e-4a38-82ec-c7256c6d6697-etc-swift\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.076919 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a525134b-d25e-4a38-82ec-c7256c6d6697-config-data\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.077650 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a525134b-d25e-4a38-82ec-c7256c6d6697-public-tls-certs\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.078305 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a525134b-d25e-4a38-82ec-c7256c6d6697-internal-tls-certs\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.080106 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a525134b-d25e-4a38-82ec-c7256c6d6697-combined-ca-bundle\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.099611 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zrhv\" (UniqueName: \"kubernetes.io/projected/a525134b-d25e-4a38-82ec-c7256c6d6697-kube-api-access-9zrhv\") pod \"swift-proxy-84b96bd867-hvzp6\" (UID: \"a525134b-d25e-4a38-82ec-c7256c6d6697\") " pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.139943 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.364969 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="954e4c03-8c79-4207-8f16-97d8db1c1b25" path="/var/lib/kubelet/pods/954e4c03-8c79-4207-8f16-97d8db1c1b25/volumes" Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.734274 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-84b96bd867-hvzp6"] Nov 24 17:44:20 crc kubenswrapper[4808]: I1124 17:44:20.862985 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 17:44:21 crc kubenswrapper[4808]: I1124 17:44:21.505844 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-84b96bd867-hvzp6" event={"ID":"a525134b-d25e-4a38-82ec-c7256c6d6697","Type":"ContainerStarted","Data":"7e78a826f2c9f031e0ace057ccee622abb828b26d88414aca6f8c917c4762628"} Nov 24 17:44:21 crc kubenswrapper[4808]: I1124 17:44:21.506223 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-84b96bd867-hvzp6" event={"ID":"a525134b-d25e-4a38-82ec-c7256c6d6697","Type":"ContainerStarted","Data":"404fb1f97b66ec653b9adb8875423716042576870dd259b81e2ea4c3b3d5613f"} Nov 24 17:44:21 crc kubenswrapper[4808]: I1124 17:44:21.506243 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-84b96bd867-hvzp6" event={"ID":"a525134b-d25e-4a38-82ec-c7256c6d6697","Type":"ContainerStarted","Data":"7e9f2473569f249f841daf4aa92710d736a22139cc780182dedb9c51ec53e9a9"} Nov 24 17:44:21 crc kubenswrapper[4808]: I1124 17:44:21.506606 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:21 crc kubenswrapper[4808]: I1124 17:44:21.506629 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:21 crc kubenswrapper[4808]: I1124 17:44:21.538576 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-84b96bd867-hvzp6" podStartSLOduration=2.538558466 podStartE2EDuration="2.538558466s" podCreationTimestamp="2025-11-24 17:44:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:44:21.529657646 +0000 UTC m=+1054.127325458" watchObservedRunningTime="2025-11-24 17:44:21.538558466 +0000 UTC m=+1054.136226268" Nov 24 17:44:21 crc kubenswrapper[4808]: I1124 17:44:21.997076 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:21 crc kubenswrapper[4808]: I1124 17:44:21.998343 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="proxy-httpd" containerID="cri-o://3cd780b72720ae5f11014a2d3ca9c7f99fd586c20e6e6c6cf2717c64ed315d1a" gracePeriod=30 Nov 24 17:44:21 crc kubenswrapper[4808]: I1124 17:44:21.998464 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="sg-core" containerID="cri-o://099ebd2a2f88f6ffb39327f15dba057a61cec83a54fe57ca5a656e1209ec47a2" gracePeriod=30 Nov 24 17:44:21 crc kubenswrapper[4808]: I1124 17:44:21.998511 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="ceilometer-notification-agent" containerID="cri-o://e23989b3dae6ccfc9f8162a8b675245071adf74a526be0447364a3ca369ae6d4" gracePeriod=30 Nov 24 17:44:21 crc kubenswrapper[4808]: I1124 17:44:21.998563 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="ceilometer-central-agent" containerID="cri-o://9469ceb2a3994298c8cec3b2a29b663e124431ebf84c778b635cda65cdc7bac1" gracePeriod=30 Nov 24 17:44:22 crc kubenswrapper[4808]: I1124 17:44:22.104748 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.166:3000/\": read tcp 10.217.0.2:46608->10.217.0.166:3000: read: connection reset by peer" Nov 24 17:44:22 crc kubenswrapper[4808]: I1124 17:44:22.520127 4808 generic.go:334] "Generic (PLEG): container finished" podID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerID="3cd780b72720ae5f11014a2d3ca9c7f99fd586c20e6e6c6cf2717c64ed315d1a" exitCode=0 Nov 24 17:44:22 crc kubenswrapper[4808]: I1124 17:44:22.520173 4808 generic.go:334] "Generic (PLEG): container finished" podID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerID="099ebd2a2f88f6ffb39327f15dba057a61cec83a54fe57ca5a656e1209ec47a2" exitCode=2 Nov 24 17:44:22 crc kubenswrapper[4808]: I1124 17:44:22.520240 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79bda77a-32e2-4bbf-a304-7e827af401e8","Type":"ContainerDied","Data":"3cd780b72720ae5f11014a2d3ca9c7f99fd586c20e6e6c6cf2717c64ed315d1a"} Nov 24 17:44:22 crc kubenswrapper[4808]: I1124 17:44:22.520311 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79bda77a-32e2-4bbf-a304-7e827af401e8","Type":"ContainerDied","Data":"099ebd2a2f88f6ffb39327f15dba057a61cec83a54fe57ca5a656e1209ec47a2"} Nov 24 17:44:23 crc kubenswrapper[4808]: I1124 17:44:23.532064 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79bda77a-32e2-4bbf-a304-7e827af401e8","Type":"ContainerDied","Data":"e23989b3dae6ccfc9f8162a8b675245071adf74a526be0447364a3ca369ae6d4"} Nov 24 17:44:23 crc kubenswrapper[4808]: I1124 17:44:23.532074 4808 generic.go:334] "Generic (PLEG): container finished" podID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerID="e23989b3dae6ccfc9f8162a8b675245071adf74a526be0447364a3ca369ae6d4" exitCode=0 Nov 24 17:44:23 crc kubenswrapper[4808]: I1124 17:44:23.532439 4808 generic.go:334] "Generic (PLEG): container finished" podID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerID="9469ceb2a3994298c8cec3b2a29b663e124431ebf84c778b635cda65cdc7bac1" exitCode=0 Nov 24 17:44:23 crc kubenswrapper[4808]: I1124 17:44:23.532465 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79bda77a-32e2-4bbf-a304-7e827af401e8","Type":"ContainerDied","Data":"9469ceb2a3994298c8cec3b2a29b663e124431ebf84c778b635cda65cdc7bac1"} Nov 24 17:44:25 crc kubenswrapper[4808]: I1124 17:44:25.155422 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:26 crc kubenswrapper[4808]: I1124 17:44:26.115589 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 17:44:26 crc kubenswrapper[4808]: I1124 17:44:26.473108 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.166:3000/\": dial tcp 10.217.0.166:3000: connect: connection refused" Nov 24 17:44:26 crc kubenswrapper[4808]: I1124 17:44:26.963372 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:44:26 crc kubenswrapper[4808]: I1124 17:44:26.963867 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="25fd4604-12b2-412f-8d76-15584feda527" containerName="glance-log" containerID="cri-o://3f2f328d9ec9832c76186617ad9dafc44adaa3d7d110568d0dbaede229b47bc3" gracePeriod=30 Nov 24 17:44:26 crc kubenswrapper[4808]: I1124 17:44:26.963927 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="25fd4604-12b2-412f-8d76-15584feda527" containerName="glance-httpd" containerID="cri-o://67c1c320b47c34a7be66808fdc9afea31d85a153ca43935389616dd85d6cb5b0" gracePeriod=30 Nov 24 17:44:27 crc kubenswrapper[4808]: I1124 17:44:27.573446 4808 generic.go:334] "Generic (PLEG): container finished" podID="25fd4604-12b2-412f-8d76-15584feda527" containerID="3f2f328d9ec9832c76186617ad9dafc44adaa3d7d110568d0dbaede229b47bc3" exitCode=143 Nov 24 17:44:27 crc kubenswrapper[4808]: I1124 17:44:27.573515 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"25fd4604-12b2-412f-8d76-15584feda527","Type":"ContainerDied","Data":"3f2f328d9ec9832c76186617ad9dafc44adaa3d7d110568d0dbaede229b47bc3"} Nov 24 17:44:27 crc kubenswrapper[4808]: I1124 17:44:27.677694 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:44:27 crc kubenswrapper[4808]: I1124 17:44:27.677907 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="21208d77-22ac-428a-a5a0-39c69fe1fa6f" containerName="glance-log" containerID="cri-o://f54aee1e3c31595f4cc7fd12294fbcf12fd01b06a02f9d9c509fe9430a81d96e" gracePeriod=30 Nov 24 17:44:27 crc kubenswrapper[4808]: I1124 17:44:27.678346 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="21208d77-22ac-428a-a5a0-39c69fe1fa6f" containerName="glance-httpd" containerID="cri-o://8996b50c2d9fcdd9da3aa3e14d475f35e97b6f081cfc97b0e348e7fb1d0972c9" gracePeriod=30 Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.172445 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.233995 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lxcp\" (UniqueName: \"kubernetes.io/projected/79bda77a-32e2-4bbf-a304-7e827af401e8-kube-api-access-7lxcp\") pod \"79bda77a-32e2-4bbf-a304-7e827af401e8\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.234077 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-sg-core-conf-yaml\") pod \"79bda77a-32e2-4bbf-a304-7e827af401e8\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.234105 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-config-data\") pod \"79bda77a-32e2-4bbf-a304-7e827af401e8\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.234125 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79bda77a-32e2-4bbf-a304-7e827af401e8-run-httpd\") pod \"79bda77a-32e2-4bbf-a304-7e827af401e8\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.234145 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-combined-ca-bundle\") pod \"79bda77a-32e2-4bbf-a304-7e827af401e8\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.234159 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79bda77a-32e2-4bbf-a304-7e827af401e8-log-httpd\") pod \"79bda77a-32e2-4bbf-a304-7e827af401e8\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.234228 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-scripts\") pod \"79bda77a-32e2-4bbf-a304-7e827af401e8\" (UID: \"79bda77a-32e2-4bbf-a304-7e827af401e8\") " Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.237854 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79bda77a-32e2-4bbf-a304-7e827af401e8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "79bda77a-32e2-4bbf-a304-7e827af401e8" (UID: "79bda77a-32e2-4bbf-a304-7e827af401e8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.237974 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79bda77a-32e2-4bbf-a304-7e827af401e8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "79bda77a-32e2-4bbf-a304-7e827af401e8" (UID: "79bda77a-32e2-4bbf-a304-7e827af401e8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.243867 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79bda77a-32e2-4bbf-a304-7e827af401e8-kube-api-access-7lxcp" (OuterVolumeSpecName: "kube-api-access-7lxcp") pod "79bda77a-32e2-4bbf-a304-7e827af401e8" (UID: "79bda77a-32e2-4bbf-a304-7e827af401e8"). InnerVolumeSpecName "kube-api-access-7lxcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.248155 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-scripts" (OuterVolumeSpecName: "scripts") pod "79bda77a-32e2-4bbf-a304-7e827af401e8" (UID: "79bda77a-32e2-4bbf-a304-7e827af401e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.264446 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "79bda77a-32e2-4bbf-a304-7e827af401e8" (UID: "79bda77a-32e2-4bbf-a304-7e827af401e8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.315648 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79bda77a-32e2-4bbf-a304-7e827af401e8" (UID: "79bda77a-32e2-4bbf-a304-7e827af401e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.336406 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lxcp\" (UniqueName: \"kubernetes.io/projected/79bda77a-32e2-4bbf-a304-7e827af401e8-kube-api-access-7lxcp\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.336482 4808 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.336497 4808 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79bda77a-32e2-4bbf-a304-7e827af401e8-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.336508 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.336520 4808 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79bda77a-32e2-4bbf-a304-7e827af401e8-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.336530 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.364741 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-config-data" (OuterVolumeSpecName: "config-data") pod "79bda77a-32e2-4bbf-a304-7e827af401e8" (UID: "79bda77a-32e2-4bbf-a304-7e827af401e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.438417 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79bda77a-32e2-4bbf-a304-7e827af401e8-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.583549 4808 generic.go:334] "Generic (PLEG): container finished" podID="21208d77-22ac-428a-a5a0-39c69fe1fa6f" containerID="f54aee1e3c31595f4cc7fd12294fbcf12fd01b06a02f9d9c509fe9430a81d96e" exitCode=143 Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.583606 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"21208d77-22ac-428a-a5a0-39c69fe1fa6f","Type":"ContainerDied","Data":"f54aee1e3c31595f4cc7fd12294fbcf12fd01b06a02f9d9c509fe9430a81d96e"} Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.584857 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c","Type":"ContainerStarted","Data":"49bad4a8d50444e7612b34b83f9a9a5d2ac02f61f03c33e212f41eb885eb671d"} Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.588612 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79bda77a-32e2-4bbf-a304-7e827af401e8","Type":"ContainerDied","Data":"3d02da0b803096c40d1d772d68cac23f8867b51a7c0b49ce7936b117bcbbe1a9"} Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.588641 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.588652 4808 scope.go:117] "RemoveContainer" containerID="3cd780b72720ae5f11014a2d3ca9c7f99fd586c20e6e6c6cf2717c64ed315d1a" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.601739 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.358955944 podStartE2EDuration="13.601721665s" podCreationTimestamp="2025-11-24 17:44:15 +0000 UTC" firstStartedPulling="2025-11-24 17:44:16.635438273 +0000 UTC m=+1049.233106075" lastFinishedPulling="2025-11-24 17:44:27.878203994 +0000 UTC m=+1060.475871796" observedRunningTime="2025-11-24 17:44:28.601081276 +0000 UTC m=+1061.198749098" watchObservedRunningTime="2025-11-24 17:44:28.601721665 +0000 UTC m=+1061.199389467" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.612717 4808 scope.go:117] "RemoveContainer" containerID="099ebd2a2f88f6ffb39327f15dba057a61cec83a54fe57ca5a656e1209ec47a2" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.628997 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.638704 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.640491 4808 scope.go:117] "RemoveContainer" containerID="e23989b3dae6ccfc9f8162a8b675245071adf74a526be0447364a3ca369ae6d4" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.650447 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:28 crc kubenswrapper[4808]: E1124 17:44:28.650989 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="ceilometer-notification-agent" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.651971 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="ceilometer-notification-agent" Nov 24 17:44:28 crc kubenswrapper[4808]: E1124 17:44:28.652066 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="sg-core" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.652119 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="sg-core" Nov 24 17:44:28 crc kubenswrapper[4808]: E1124 17:44:28.652188 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="ceilometer-central-agent" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.652239 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="ceilometer-central-agent" Nov 24 17:44:28 crc kubenswrapper[4808]: E1124 17:44:28.652351 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="proxy-httpd" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.652408 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="proxy-httpd" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.652659 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="ceilometer-notification-agent" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.652727 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="ceilometer-central-agent" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.652809 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="proxy-httpd" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.652882 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" containerName="sg-core" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.654778 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.660260 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.668356 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.677812 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.701380 4808 scope.go:117] "RemoveContainer" containerID="9469ceb2a3994298c8cec3b2a29b663e124431ebf84c778b635cda65cdc7bac1" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.743484 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.743584 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-scripts\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.743674 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.743740 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-config-data\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.743792 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17d53839-6156-4d7e-ad19-b8a092398cf6-run-httpd\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.743812 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17d53839-6156-4d7e-ad19-b8a092398cf6-log-httpd\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.743935 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwwg2\" (UniqueName: \"kubernetes.io/projected/17d53839-6156-4d7e-ad19-b8a092398cf6-kube-api-access-mwwg2\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.846455 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17d53839-6156-4d7e-ad19-b8a092398cf6-run-httpd\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.846749 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17d53839-6156-4d7e-ad19-b8a092398cf6-log-httpd\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.846865 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwwg2\" (UniqueName: \"kubernetes.io/projected/17d53839-6156-4d7e-ad19-b8a092398cf6-kube-api-access-mwwg2\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.846958 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.847073 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-scripts\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.846961 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17d53839-6156-4d7e-ad19-b8a092398cf6-run-httpd\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.847247 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17d53839-6156-4d7e-ad19-b8a092398cf6-log-httpd\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.847376 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.847491 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-config-data\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.850908 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.851079 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-scripts\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.851128 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-config-data\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.851899 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:28 crc kubenswrapper[4808]: I1124 17:44:28.864506 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwwg2\" (UniqueName: \"kubernetes.io/projected/17d53839-6156-4d7e-ad19-b8a092398cf6-kube-api-access-mwwg2\") pod \"ceilometer-0\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " pod="openstack/ceilometer-0" Nov 24 17:44:29 crc kubenswrapper[4808]: I1124 17:44:29.015886 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:44:29 crc kubenswrapper[4808]: I1124 17:44:29.467278 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:29 crc kubenswrapper[4808]: W1124 17:44:29.469940 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17d53839_6156_4d7e_ad19_b8a092398cf6.slice/crio-5e47702aa44cc848afe866721afb95ff31456a8c2f74d9d9041a15ced65976be WatchSource:0}: Error finding container 5e47702aa44cc848afe866721afb95ff31456a8c2f74d9d9041a15ced65976be: Status 404 returned error can't find the container with id 5e47702aa44cc848afe866721afb95ff31456a8c2f74d9d9041a15ced65976be Nov 24 17:44:29 crc kubenswrapper[4808]: I1124 17:44:29.598620 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17d53839-6156-4d7e-ad19-b8a092398cf6","Type":"ContainerStarted","Data":"5e47702aa44cc848afe866721afb95ff31456a8c2f74d9d9041a15ced65976be"} Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.143420 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-dvnqp"] Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.145973 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-dvnqp" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.159738 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-dvnqp"] Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.170790 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-84b96bd867-hvzp6" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.196567 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b835d11-9481-4841-8a77-cc923fbb999b-operator-scripts\") pod \"nova-api-db-create-dvnqp\" (UID: \"5b835d11-9481-4841-8a77-cc923fbb999b\") " pod="openstack/nova-api-db-create-dvnqp" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.196676 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2txw\" (UniqueName: \"kubernetes.io/projected/5b835d11-9481-4841-8a77-cc923fbb999b-kube-api-access-s2txw\") pod \"nova-api-db-create-dvnqp\" (UID: \"5b835d11-9481-4841-8a77-cc923fbb999b\") " pod="openstack/nova-api-db-create-dvnqp" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.280917 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-7czsx"] Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.282278 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7czsx" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.298505 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b835d11-9481-4841-8a77-cc923fbb999b-operator-scripts\") pod \"nova-api-db-create-dvnqp\" (UID: \"5b835d11-9481-4841-8a77-cc923fbb999b\") " pod="openstack/nova-api-db-create-dvnqp" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.298654 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2txw\" (UniqueName: \"kubernetes.io/projected/5b835d11-9481-4841-8a77-cc923fbb999b-kube-api-access-s2txw\") pod \"nova-api-db-create-dvnqp\" (UID: \"5b835d11-9481-4841-8a77-cc923fbb999b\") " pod="openstack/nova-api-db-create-dvnqp" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.300451 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b835d11-9481-4841-8a77-cc923fbb999b-operator-scripts\") pod \"nova-api-db-create-dvnqp\" (UID: \"5b835d11-9481-4841-8a77-cc923fbb999b\") " pod="openstack/nova-api-db-create-dvnqp" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.314148 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-7czsx"] Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.334830 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2txw\" (UniqueName: \"kubernetes.io/projected/5b835d11-9481-4841-8a77-cc923fbb999b-kube-api-access-s2txw\") pod \"nova-api-db-create-dvnqp\" (UID: \"5b835d11-9481-4841-8a77-cc923fbb999b\") " pod="openstack/nova-api-db-create-dvnqp" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.391693 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79bda77a-32e2-4bbf-a304-7e827af401e8" path="/var/lib/kubelet/pods/79bda77a-32e2-4bbf-a304-7e827af401e8/volumes" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.392413 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-6nxpn"] Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.393401 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-cec5-account-create-update-vzfxt"] Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.394326 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6nxpn" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.395607 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-6nxpn"] Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.395684 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-cec5-account-create-update-vzfxt" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.396545 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-cec5-account-create-update-vzfxt"] Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.397255 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.400049 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d-operator-scripts\") pod \"nova-cell0-db-create-7czsx\" (UID: \"09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d\") " pod="openstack/nova-cell0-db-create-7czsx" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.400085 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8xjf\" (UniqueName: \"kubernetes.io/projected/09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d-kube-api-access-c8xjf\") pod \"nova-cell0-db-create-7czsx\" (UID: \"09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d\") " pod="openstack/nova-cell0-db-create-7czsx" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.477179 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-dvnqp" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.501994 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lppwk\" (UniqueName: \"kubernetes.io/projected/76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c-kube-api-access-lppwk\") pod \"nova-api-cec5-account-create-update-vzfxt\" (UID: \"76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c\") " pod="openstack/nova-api-cec5-account-create-update-vzfxt" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.502061 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cpdp\" (UniqueName: \"kubernetes.io/projected/0a398e94-3f7a-48ec-815f-79c5a095c9c6-kube-api-access-2cpdp\") pod \"nova-cell1-db-create-6nxpn\" (UID: \"0a398e94-3f7a-48ec-815f-79c5a095c9c6\") " pod="openstack/nova-cell1-db-create-6nxpn" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.502087 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c-operator-scripts\") pod \"nova-api-cec5-account-create-update-vzfxt\" (UID: \"76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c\") " pod="openstack/nova-api-cec5-account-create-update-vzfxt" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.502138 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a398e94-3f7a-48ec-815f-79c5a095c9c6-operator-scripts\") pod \"nova-cell1-db-create-6nxpn\" (UID: \"0a398e94-3f7a-48ec-815f-79c5a095c9c6\") " pod="openstack/nova-cell1-db-create-6nxpn" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.502229 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d-operator-scripts\") pod \"nova-cell0-db-create-7czsx\" (UID: \"09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d\") " pod="openstack/nova-cell0-db-create-7czsx" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.502520 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8xjf\" (UniqueName: \"kubernetes.io/projected/09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d-kube-api-access-c8xjf\") pod \"nova-cell0-db-create-7czsx\" (UID: \"09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d\") " pod="openstack/nova-cell0-db-create-7czsx" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.503139 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d-operator-scripts\") pod \"nova-cell0-db-create-7czsx\" (UID: \"09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d\") " pod="openstack/nova-cell0-db-create-7czsx" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.525822 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8xjf\" (UniqueName: \"kubernetes.io/projected/09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d-kube-api-access-c8xjf\") pod \"nova-cell0-db-create-7czsx\" (UID: \"09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d\") " pod="openstack/nova-cell0-db-create-7czsx" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.554903 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-f761-account-create-update-dmmw8"] Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.558100 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f761-account-create-update-dmmw8" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.562866 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.581414 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f761-account-create-update-dmmw8"] Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.603579 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cpdp\" (UniqueName: \"kubernetes.io/projected/0a398e94-3f7a-48ec-815f-79c5a095c9c6-kube-api-access-2cpdp\") pod \"nova-cell1-db-create-6nxpn\" (UID: \"0a398e94-3f7a-48ec-815f-79c5a095c9c6\") " pod="openstack/nova-cell1-db-create-6nxpn" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.603634 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c-operator-scripts\") pod \"nova-api-cec5-account-create-update-vzfxt\" (UID: \"76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c\") " pod="openstack/nova-api-cec5-account-create-update-vzfxt" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.603709 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a398e94-3f7a-48ec-815f-79c5a095c9c6-operator-scripts\") pod \"nova-cell1-db-create-6nxpn\" (UID: \"0a398e94-3f7a-48ec-815f-79c5a095c9c6\") " pod="openstack/nova-cell1-db-create-6nxpn" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.604462 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szwr2\" (UniqueName: \"kubernetes.io/projected/87188566-f3cf-4cea-bdd1-28f3c48f5b37-kube-api-access-szwr2\") pod \"nova-cell0-f761-account-create-update-dmmw8\" (UID: \"87188566-f3cf-4cea-bdd1-28f3c48f5b37\") " pod="openstack/nova-cell0-f761-account-create-update-dmmw8" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.604675 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87188566-f3cf-4cea-bdd1-28f3c48f5b37-operator-scripts\") pod \"nova-cell0-f761-account-create-update-dmmw8\" (UID: \"87188566-f3cf-4cea-bdd1-28f3c48f5b37\") " pod="openstack/nova-cell0-f761-account-create-update-dmmw8" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.604778 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lppwk\" (UniqueName: \"kubernetes.io/projected/76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c-kube-api-access-lppwk\") pod \"nova-api-cec5-account-create-update-vzfxt\" (UID: \"76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c\") " pod="openstack/nova-api-cec5-account-create-update-vzfxt" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.605159 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c-operator-scripts\") pod \"nova-api-cec5-account-create-update-vzfxt\" (UID: \"76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c\") " pod="openstack/nova-api-cec5-account-create-update-vzfxt" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.609259 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a398e94-3f7a-48ec-815f-79c5a095c9c6-operator-scripts\") pod \"nova-cell1-db-create-6nxpn\" (UID: \"0a398e94-3f7a-48ec-815f-79c5a095c9c6\") " pod="openstack/nova-cell1-db-create-6nxpn" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.625928 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cpdp\" (UniqueName: \"kubernetes.io/projected/0a398e94-3f7a-48ec-815f-79c5a095c9c6-kube-api-access-2cpdp\") pod \"nova-cell1-db-create-6nxpn\" (UID: \"0a398e94-3f7a-48ec-815f-79c5a095c9c6\") " pod="openstack/nova-cell1-db-create-6nxpn" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.640681 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lppwk\" (UniqueName: \"kubernetes.io/projected/76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c-kube-api-access-lppwk\") pod \"nova-api-cec5-account-create-update-vzfxt\" (UID: \"76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c\") " pod="openstack/nova-api-cec5-account-create-update-vzfxt" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.645460 4808 generic.go:334] "Generic (PLEG): container finished" podID="25fd4604-12b2-412f-8d76-15584feda527" containerID="67c1c320b47c34a7be66808fdc9afea31d85a153ca43935389616dd85d6cb5b0" exitCode=0 Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.645541 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"25fd4604-12b2-412f-8d76-15584feda527","Type":"ContainerDied","Data":"67c1c320b47c34a7be66808fdc9afea31d85a153ca43935389616dd85d6cb5b0"} Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.651297 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17d53839-6156-4d7e-ad19-b8a092398cf6","Type":"ContainerStarted","Data":"f4141ea6636038d9a14bd6b74666b2906d67f622759fc2868403445ac565f3d3"} Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.659049 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7czsx" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.707697 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szwr2\" (UniqueName: \"kubernetes.io/projected/87188566-f3cf-4cea-bdd1-28f3c48f5b37-kube-api-access-szwr2\") pod \"nova-cell0-f761-account-create-update-dmmw8\" (UID: \"87188566-f3cf-4cea-bdd1-28f3c48f5b37\") " pod="openstack/nova-cell0-f761-account-create-update-dmmw8" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.707873 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87188566-f3cf-4cea-bdd1-28f3c48f5b37-operator-scripts\") pod \"nova-cell0-f761-account-create-update-dmmw8\" (UID: \"87188566-f3cf-4cea-bdd1-28f3c48f5b37\") " pod="openstack/nova-cell0-f761-account-create-update-dmmw8" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.708763 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87188566-f3cf-4cea-bdd1-28f3c48f5b37-operator-scripts\") pod \"nova-cell0-f761-account-create-update-dmmw8\" (UID: \"87188566-f3cf-4cea-bdd1-28f3c48f5b37\") " pod="openstack/nova-cell0-f761-account-create-update-dmmw8" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.728368 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6nxpn" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.738630 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szwr2\" (UniqueName: \"kubernetes.io/projected/87188566-f3cf-4cea-bdd1-28f3c48f5b37-kube-api-access-szwr2\") pod \"nova-cell0-f761-account-create-update-dmmw8\" (UID: \"87188566-f3cf-4cea-bdd1-28f3c48f5b37\") " pod="openstack/nova-cell0-f761-account-create-update-dmmw8" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.744914 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-cec5-account-create-update-vzfxt" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.777195 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-4a1f-account-create-update-tp9xc"] Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.780893 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4a1f-account-create-update-tp9xc" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.785166 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.810926 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/752ed614-a980-483f-a0ea-6fd150a6b694-operator-scripts\") pod \"nova-cell1-4a1f-account-create-update-tp9xc\" (UID: \"752ed614-a980-483f-a0ea-6fd150a6b694\") " pod="openstack/nova-cell1-4a1f-account-create-update-tp9xc" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.811343 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vslv9\" (UniqueName: \"kubernetes.io/projected/752ed614-a980-483f-a0ea-6fd150a6b694-kube-api-access-vslv9\") pod \"nova-cell1-4a1f-account-create-update-tp9xc\" (UID: \"752ed614-a980-483f-a0ea-6fd150a6b694\") " pod="openstack/nova-cell1-4a1f-account-create-update-tp9xc" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.836301 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4a1f-account-create-update-tp9xc"] Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.889418 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f761-account-create-update-dmmw8" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.912952 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vslv9\" (UniqueName: \"kubernetes.io/projected/752ed614-a980-483f-a0ea-6fd150a6b694-kube-api-access-vslv9\") pod \"nova-cell1-4a1f-account-create-update-tp9xc\" (UID: \"752ed614-a980-483f-a0ea-6fd150a6b694\") " pod="openstack/nova-cell1-4a1f-account-create-update-tp9xc" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.913069 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/752ed614-a980-483f-a0ea-6fd150a6b694-operator-scripts\") pod \"nova-cell1-4a1f-account-create-update-tp9xc\" (UID: \"752ed614-a980-483f-a0ea-6fd150a6b694\") " pod="openstack/nova-cell1-4a1f-account-create-update-tp9xc" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.916639 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/752ed614-a980-483f-a0ea-6fd150a6b694-operator-scripts\") pod \"nova-cell1-4a1f-account-create-update-tp9xc\" (UID: \"752ed614-a980-483f-a0ea-6fd150a6b694\") " pod="openstack/nova-cell1-4a1f-account-create-update-tp9xc" Nov 24 17:44:30 crc kubenswrapper[4808]: I1124 17:44:30.948668 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vslv9\" (UniqueName: \"kubernetes.io/projected/752ed614-a980-483f-a0ea-6fd150a6b694-kube-api-access-vslv9\") pod \"nova-cell1-4a1f-account-create-update-tp9xc\" (UID: \"752ed614-a980-483f-a0ea-6fd150a6b694\") " pod="openstack/nova-cell1-4a1f-account-create-update-tp9xc" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.048742 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-dvnqp"] Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.219532 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4a1f-account-create-update-tp9xc" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.229137 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.333124 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25fd4604-12b2-412f-8d76-15584feda527-httpd-run\") pod \"25fd4604-12b2-412f-8d76-15584feda527\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.333231 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-config-data\") pod \"25fd4604-12b2-412f-8d76-15584feda527\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.333262 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25fd4604-12b2-412f-8d76-15584feda527-logs\") pod \"25fd4604-12b2-412f-8d76-15584feda527\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.333284 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-scripts\") pod \"25fd4604-12b2-412f-8d76-15584feda527\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.333309 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"25fd4604-12b2-412f-8d76-15584feda527\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.333389 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjzbr\" (UniqueName: \"kubernetes.io/projected/25fd4604-12b2-412f-8d76-15584feda527-kube-api-access-gjzbr\") pod \"25fd4604-12b2-412f-8d76-15584feda527\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.333414 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-public-tls-certs\") pod \"25fd4604-12b2-412f-8d76-15584feda527\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.333490 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-combined-ca-bundle\") pod \"25fd4604-12b2-412f-8d76-15584feda527\" (UID: \"25fd4604-12b2-412f-8d76-15584feda527\") " Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.348620 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25fd4604-12b2-412f-8d76-15584feda527-logs" (OuterVolumeSpecName: "logs") pod "25fd4604-12b2-412f-8d76-15584feda527" (UID: "25fd4604-12b2-412f-8d76-15584feda527"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.348998 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25fd4604-12b2-412f-8d76-15584feda527-kube-api-access-gjzbr" (OuterVolumeSpecName: "kube-api-access-gjzbr") pod "25fd4604-12b2-412f-8d76-15584feda527" (UID: "25fd4604-12b2-412f-8d76-15584feda527"). InnerVolumeSpecName "kube-api-access-gjzbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.350650 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25fd4604-12b2-412f-8d76-15584feda527-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "25fd4604-12b2-412f-8d76-15584feda527" (UID: "25fd4604-12b2-412f-8d76-15584feda527"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.351117 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "25fd4604-12b2-412f-8d76-15584feda527" (UID: "25fd4604-12b2-412f-8d76-15584feda527"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.368540 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-scripts" (OuterVolumeSpecName: "scripts") pod "25fd4604-12b2-412f-8d76-15584feda527" (UID: "25fd4604-12b2-412f-8d76-15584feda527"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.385856 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-7czsx"] Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.409454 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25fd4604-12b2-412f-8d76-15584feda527" (UID: "25fd4604-12b2-412f-8d76-15584feda527"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.437957 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjzbr\" (UniqueName: \"kubernetes.io/projected/25fd4604-12b2-412f-8d76-15584feda527-kube-api-access-gjzbr\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.437979 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.437988 4808 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25fd4604-12b2-412f-8d76-15584feda527-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.437996 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.438005 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25fd4604-12b2-412f-8d76-15584feda527-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.438034 4808 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.456322 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "25fd4604-12b2-412f-8d76-15584feda527" (UID: "25fd4604-12b2-412f-8d76-15584feda527"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.487290 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-6nxpn"] Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.506170 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-config-data" (OuterVolumeSpecName: "config-data") pod "25fd4604-12b2-412f-8d76-15584feda527" (UID: "25fd4604-12b2-412f-8d76-15584feda527"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.514485 4808 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.541298 4808 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.541332 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25fd4604-12b2-412f-8d76-15584feda527-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.541344 4808 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.610177 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f761-account-create-update-dmmw8"] Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.624995 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-cec5-account-create-update-vzfxt"] Nov 24 17:44:31 crc kubenswrapper[4808]: W1124 17:44:31.639217 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87188566_f3cf_4cea_bdd1_28f3c48f5b37.slice/crio-b71d42a3470b120b418e53a61227264d1b6747ff7e6250de550a62d0d7b09a21 WatchSource:0}: Error finding container b71d42a3470b120b418e53a61227264d1b6747ff7e6250de550a62d0d7b09a21: Status 404 returned error can't find the container with id b71d42a3470b120b418e53a61227264d1b6747ff7e6250de550a62d0d7b09a21 Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.685477 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"25fd4604-12b2-412f-8d76-15584feda527","Type":"ContainerDied","Data":"684be5c0677e8b973085e0d12298e58f1d12a90cf8a17a56b199dd6bccb248fe"} Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.685536 4808 scope.go:117] "RemoveContainer" containerID="67c1c320b47c34a7be66808fdc9afea31d85a153ca43935389616dd85d6cb5b0" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.685701 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.698312 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-dvnqp" event={"ID":"5b835d11-9481-4841-8a77-cc923fbb999b","Type":"ContainerStarted","Data":"e194de5e57855ecaf0a3f745c556aa64784576d24d1afb02205f8bb0d24a7d58"} Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.704291 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17d53839-6156-4d7e-ad19-b8a092398cf6","Type":"ContainerStarted","Data":"ae8ce142e42eab69b11061be4c8174ce4dc897d41fb87e7aa1e6317952b8d09a"} Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.729469 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f761-account-create-update-dmmw8" event={"ID":"87188566-f3cf-4cea-bdd1-28f3c48f5b37","Type":"ContainerStarted","Data":"b71d42a3470b120b418e53a61227264d1b6747ff7e6250de550a62d0d7b09a21"} Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.737176 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-7czsx" event={"ID":"09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d","Type":"ContainerStarted","Data":"47489a74be2e33f1acec36e8fe236864596b87db12d6939e681e2b67d70aa310"} Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.739260 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.753317 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-cec5-account-create-update-vzfxt" event={"ID":"76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c","Type":"ContainerStarted","Data":"8e45f0b2368e00215e50f9ca576e984b6ee9eefb73313a1cb3b15d0e1c465db2"} Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.756451 4808 scope.go:117] "RemoveContainer" containerID="3f2f328d9ec9832c76186617ad9dafc44adaa3d7d110568d0dbaede229b47bc3" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.771816 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.781919 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:44:31 crc kubenswrapper[4808]: E1124 17:44:31.782506 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25fd4604-12b2-412f-8d76-15584feda527" containerName="glance-httpd" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.782532 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="25fd4604-12b2-412f-8d76-15584feda527" containerName="glance-httpd" Nov 24 17:44:31 crc kubenswrapper[4808]: E1124 17:44:31.782557 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25fd4604-12b2-412f-8d76-15584feda527" containerName="glance-log" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.782566 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="25fd4604-12b2-412f-8d76-15584feda527" containerName="glance-log" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.782849 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="25fd4604-12b2-412f-8d76-15584feda527" containerName="glance-log" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.782884 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="25fd4604-12b2-412f-8d76-15584feda527" containerName="glance-httpd" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.784171 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.795260 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.797264 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.797467 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.837681 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6nxpn" event={"ID":"0a398e94-3f7a-48ec-815f-79c5a095c9c6","Type":"ContainerStarted","Data":"7bfde055b414be05268fc89772325f6880c62de48bf86f4af3c1297f2db3c33d"} Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.843457 4808 generic.go:334] "Generic (PLEG): container finished" podID="21208d77-22ac-428a-a5a0-39c69fe1fa6f" containerID="8996b50c2d9fcdd9da3aa3e14d475f35e97b6f081cfc97b0e348e7fb1d0972c9" exitCode=0 Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.843519 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"21208d77-22ac-428a-a5a0-39c69fe1fa6f","Type":"ContainerDied","Data":"8996b50c2d9fcdd9da3aa3e14d475f35e97b6f081cfc97b0e348e7fb1d0972c9"} Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.954637 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.954692 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmxz2\" (UniqueName: \"kubernetes.io/projected/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-kube-api-access-nmxz2\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.954726 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-logs\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.954745 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-scripts\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.957173 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.957203 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4a1f-account-create-update-tp9xc"] Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.957391 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-config-data\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.957434 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:31 crc kubenswrapper[4808]: I1124 17:44:31.957550 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.059645 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.059718 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmxz2\" (UniqueName: \"kubernetes.io/projected/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-kube-api-access-nmxz2\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.059757 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-logs\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.059780 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-scripts\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.059826 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.059867 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-config-data\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.059890 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.059953 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.060369 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-logs\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.060650 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.063058 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.072809 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.073316 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-config-data\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.091126 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.091009 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-scripts\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.091747 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmxz2\" (UniqueName: \"kubernetes.io/projected/dc25c6d7-a302-4ea0-bd25-2b19f9b17a71-kube-api-access-nmxz2\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.127466 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71\") " pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.207245 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.210211 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.369785 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-combined-ca-bundle\") pod \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.370717 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.370824 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-scripts\") pod \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.370846 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21208d77-22ac-428a-a5a0-39c69fe1fa6f-logs\") pod \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.370898 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58wdw\" (UniqueName: \"kubernetes.io/projected/21208d77-22ac-428a-a5a0-39c69fe1fa6f-kube-api-access-58wdw\") pod \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.370936 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21208d77-22ac-428a-a5a0-39c69fe1fa6f-httpd-run\") pod \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.370993 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-internal-tls-certs\") pod \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.371103 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-config-data\") pod \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\" (UID: \"21208d77-22ac-428a-a5a0-39c69fe1fa6f\") " Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.377128 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21208d77-22ac-428a-a5a0-39c69fe1fa6f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "21208d77-22ac-428a-a5a0-39c69fe1fa6f" (UID: "21208d77-22ac-428a-a5a0-39c69fe1fa6f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.378205 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21208d77-22ac-428a-a5a0-39c69fe1fa6f-logs" (OuterVolumeSpecName: "logs") pod "21208d77-22ac-428a-a5a0-39c69fe1fa6f" (UID: "21208d77-22ac-428a-a5a0-39c69fe1fa6f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.404570 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-scripts" (OuterVolumeSpecName: "scripts") pod "21208d77-22ac-428a-a5a0-39c69fe1fa6f" (UID: "21208d77-22ac-428a-a5a0-39c69fe1fa6f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.404754 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21208d77-22ac-428a-a5a0-39c69fe1fa6f-kube-api-access-58wdw" (OuterVolumeSpecName: "kube-api-access-58wdw") pod "21208d77-22ac-428a-a5a0-39c69fe1fa6f" (UID: "21208d77-22ac-428a-a5a0-39c69fe1fa6f"). InnerVolumeSpecName "kube-api-access-58wdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.404807 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "21208d77-22ac-428a-a5a0-39c69fe1fa6f" (UID: "21208d77-22ac-428a-a5a0-39c69fe1fa6f"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.411197 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25fd4604-12b2-412f-8d76-15584feda527" path="/var/lib/kubelet/pods/25fd4604-12b2-412f-8d76-15584feda527/volumes" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.473931 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58wdw\" (UniqueName: \"kubernetes.io/projected/21208d77-22ac-428a-a5a0-39c69fe1fa6f-kube-api-access-58wdw\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.475973 4808 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21208d77-22ac-428a-a5a0-39c69fe1fa6f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.476211 4808 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.476333 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.476413 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21208d77-22ac-428a-a5a0-39c69fe1fa6f-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.510360 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21208d77-22ac-428a-a5a0-39c69fe1fa6f" (UID: "21208d77-22ac-428a-a5a0-39c69fe1fa6f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.523371 4808 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.534215 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-config-data" (OuterVolumeSpecName: "config-data") pod "21208d77-22ac-428a-a5a0-39c69fe1fa6f" (UID: "21208d77-22ac-428a-a5a0-39c69fe1fa6f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.539120 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "21208d77-22ac-428a-a5a0-39c69fe1fa6f" (UID: "21208d77-22ac-428a-a5a0-39c69fe1fa6f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.583120 4808 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.583145 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.583154 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21208d77-22ac-428a-a5a0-39c69fe1fa6f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.583162 4808 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.851270 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.882099 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b835d11-9481-4841-8a77-cc923fbb999b" containerID="ae616192db09d828ccb1875aedcd69bcdfa2d4db65ac1018f3364bf456a209f3" exitCode=0 Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.882188 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-dvnqp" event={"ID":"5b835d11-9481-4841-8a77-cc923fbb999b","Type":"ContainerDied","Data":"ae616192db09d828ccb1875aedcd69bcdfa2d4db65ac1018f3364bf456a209f3"} Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.901469 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17d53839-6156-4d7e-ad19-b8a092398cf6","Type":"ContainerStarted","Data":"479c76abd79606ada4f9033e0dd9ff1fdffd1637e35428609952b3ddb85d3b60"} Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.911721 4808 generic.go:334] "Generic (PLEG): container finished" podID="752ed614-a980-483f-a0ea-6fd150a6b694" containerID="54379e86098921409fbbfccf84f368209dc1839fd7b8b1db807667415cb6cb70" exitCode=0 Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.911813 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4a1f-account-create-update-tp9xc" event={"ID":"752ed614-a980-483f-a0ea-6fd150a6b694","Type":"ContainerDied","Data":"54379e86098921409fbbfccf84f368209dc1839fd7b8b1db807667415cb6cb70"} Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.911839 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4a1f-account-create-update-tp9xc" event={"ID":"752ed614-a980-483f-a0ea-6fd150a6b694","Type":"ContainerStarted","Data":"1e2a8215bdb935ebf2c664cc0dd23a1c7f125be5044a587bb13298fd9deeeb59"} Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.916560 4808 generic.go:334] "Generic (PLEG): container finished" podID="09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d" containerID="8172fc826925b3aeec00348cad371bdd3ffb29488b839c7612d3b7c5c861b9ca" exitCode=0 Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.916614 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-7czsx" event={"ID":"09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d","Type":"ContainerDied","Data":"8172fc826925b3aeec00348cad371bdd3ffb29488b839c7612d3b7c5c861b9ca"} Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.918217 4808 generic.go:334] "Generic (PLEG): container finished" podID="0a398e94-3f7a-48ec-815f-79c5a095c9c6" containerID="31bad8802492f6658a7fae1e789c7fbf7adf4204d366956ca2ee46c406b4b28c" exitCode=0 Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.918260 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6nxpn" event={"ID":"0a398e94-3f7a-48ec-815f-79c5a095c9c6","Type":"ContainerDied","Data":"31bad8802492f6658a7fae1e789c7fbf7adf4204d366956ca2ee46c406b4b28c"} Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.921443 4808 generic.go:334] "Generic (PLEG): container finished" podID="87188566-f3cf-4cea-bdd1-28f3c48f5b37" containerID="b6f3277bc56acbbf522afb17fc0bf4ce8d53aed1e4d1a2fa1a244aa5d87517dc" exitCode=0 Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.921484 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f761-account-create-update-dmmw8" event={"ID":"87188566-f3cf-4cea-bdd1-28f3c48f5b37","Type":"ContainerDied","Data":"b6f3277bc56acbbf522afb17fc0bf4ce8d53aed1e4d1a2fa1a244aa5d87517dc"} Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.923107 4808 generic.go:334] "Generic (PLEG): container finished" podID="76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c" containerID="7f46ba68d7c49118fb0b987a7944bc491a47727d684d6ddf806bbe1d7562b2e6" exitCode=0 Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.923144 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-cec5-account-create-update-vzfxt" event={"ID":"76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c","Type":"ContainerDied","Data":"7f46ba68d7c49118fb0b987a7944bc491a47727d684d6ddf806bbe1d7562b2e6"} Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.924656 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"21208d77-22ac-428a-a5a0-39c69fe1fa6f","Type":"ContainerDied","Data":"d2555b21a92578f9090a7c0f0753a046ea50529b6e2084bfec5bdda50d586cae"} Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.924685 4808 scope.go:117] "RemoveContainer" containerID="8996b50c2d9fcdd9da3aa3e14d475f35e97b6f081cfc97b0e348e7fb1d0972c9" Nov 24 17:44:32 crc kubenswrapper[4808]: I1124 17:44:32.924797 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.162751 4808 scope.go:117] "RemoveContainer" containerID="f54aee1e3c31595f4cc7fd12294fbcf12fd01b06a02f9d9c509fe9430a81d96e" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.163611 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.181085 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.191261 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:44:33 crc kubenswrapper[4808]: E1124 17:44:33.191683 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21208d77-22ac-428a-a5a0-39c69fe1fa6f" containerName="glance-log" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.191704 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="21208d77-22ac-428a-a5a0-39c69fe1fa6f" containerName="glance-log" Nov 24 17:44:33 crc kubenswrapper[4808]: E1124 17:44:33.191754 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21208d77-22ac-428a-a5a0-39c69fe1fa6f" containerName="glance-httpd" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.191764 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="21208d77-22ac-428a-a5a0-39c69fe1fa6f" containerName="glance-httpd" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.191988 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="21208d77-22ac-428a-a5a0-39c69fe1fa6f" containerName="glance-httpd" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.192007 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="21208d77-22ac-428a-a5a0-39c69fe1fa6f" containerName="glance-log" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.193261 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.195365 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.199159 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.207866 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.307856 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88cb8610-eb5e-45de-8028-37fc361beaac-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.307915 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-559lc\" (UniqueName: \"kubernetes.io/projected/88cb8610-eb5e-45de-8028-37fc361beaac-kube-api-access-559lc\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.307958 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88cb8610-eb5e-45de-8028-37fc361beaac-scripts\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.307987 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88cb8610-eb5e-45de-8028-37fc361beaac-config-data\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.308134 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88cb8610-eb5e-45de-8028-37fc361beaac-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.308164 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88cb8610-eb5e-45de-8028-37fc361beaac-logs\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.308182 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/88cb8610-eb5e-45de-8028-37fc361beaac-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.308214 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.365224 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.410140 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88cb8610-eb5e-45de-8028-37fc361beaac-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.410223 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88cb8610-eb5e-45de-8028-37fc361beaac-logs\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.410272 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/88cb8610-eb5e-45de-8028-37fc361beaac-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.410319 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.410636 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.410761 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88cb8610-eb5e-45de-8028-37fc361beaac-logs\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.416911 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/88cb8610-eb5e-45de-8028-37fc361beaac-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.420788 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88cb8610-eb5e-45de-8028-37fc361beaac-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.430335 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88cb8610-eb5e-45de-8028-37fc361beaac-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.430467 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-559lc\" (UniqueName: \"kubernetes.io/projected/88cb8610-eb5e-45de-8028-37fc361beaac-kube-api-access-559lc\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.430545 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88cb8610-eb5e-45de-8028-37fc361beaac-scripts\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.430569 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88cb8610-eb5e-45de-8028-37fc361beaac-config-data\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.431658 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88cb8610-eb5e-45de-8028-37fc361beaac-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.447791 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88cb8610-eb5e-45de-8028-37fc361beaac-scripts\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.450434 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.451998 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-559lc\" (UniqueName: \"kubernetes.io/projected/88cb8610-eb5e-45de-8028-37fc361beaac-kube-api-access-559lc\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.452511 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88cb8610-eb5e-45de-8028-37fc361beaac-config-data\") pod \"glance-default-internal-api-0\" (UID: \"88cb8610-eb5e-45de-8028-37fc361beaac\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.511468 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.953767 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71","Type":"ContainerStarted","Data":"4a4199052f304bdc8cf0b1b8751841f2ba777cfade5a7f410b6a853c9a1a8821"} Nov 24 17:44:33 crc kubenswrapper[4808]: I1124 17:44:33.954089 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71","Type":"ContainerStarted","Data":"420cee33e8f8bd87311ee65ba88e85d1a6d5f5a9935421421fe93b8f2ee3e6e9"} Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.142822 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.391753 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21208d77-22ac-428a-a5a0-39c69fe1fa6f" path="/var/lib/kubelet/pods/21208d77-22ac-428a-a5a0-39c69fe1fa6f/volumes" Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.503744 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-cec5-account-create-update-vzfxt" Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.570282 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6nxpn" Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.653864 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cpdp\" (UniqueName: \"kubernetes.io/projected/0a398e94-3f7a-48ec-815f-79c5a095c9c6-kube-api-access-2cpdp\") pod \"0a398e94-3f7a-48ec-815f-79c5a095c9c6\" (UID: \"0a398e94-3f7a-48ec-815f-79c5a095c9c6\") " Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.653922 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lppwk\" (UniqueName: \"kubernetes.io/projected/76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c-kube-api-access-lppwk\") pod \"76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c\" (UID: \"76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c\") " Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.653993 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a398e94-3f7a-48ec-815f-79c5a095c9c6-operator-scripts\") pod \"0a398e94-3f7a-48ec-815f-79c5a095c9c6\" (UID: \"0a398e94-3f7a-48ec-815f-79c5a095c9c6\") " Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.654092 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c-operator-scripts\") pod \"76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c\" (UID: \"76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c\") " Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.654847 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c" (UID: "76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.654854 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a398e94-3f7a-48ec-815f-79c5a095c9c6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0a398e94-3f7a-48ec-815f-79c5a095c9c6" (UID: "0a398e94-3f7a-48ec-815f-79c5a095c9c6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.660356 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a398e94-3f7a-48ec-815f-79c5a095c9c6-kube-api-access-2cpdp" (OuterVolumeSpecName: "kube-api-access-2cpdp") pod "0a398e94-3f7a-48ec-815f-79c5a095c9c6" (UID: "0a398e94-3f7a-48ec-815f-79c5a095c9c6"). InnerVolumeSpecName "kube-api-access-2cpdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.664305 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c-kube-api-access-lppwk" (OuterVolumeSpecName: "kube-api-access-lppwk") pod "76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c" (UID: "76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c"). InnerVolumeSpecName "kube-api-access-lppwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.756353 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a398e94-3f7a-48ec-815f-79c5a095c9c6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.756377 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.756390 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cpdp\" (UniqueName: \"kubernetes.io/projected/0a398e94-3f7a-48ec-815f-79c5a095c9c6-kube-api-access-2cpdp\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.756408 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lppwk\" (UniqueName: \"kubernetes.io/projected/76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c-kube-api-access-lppwk\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.976858 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-dvnqp" Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.980693 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4a1f-account-create-update-tp9xc" event={"ID":"752ed614-a980-483f-a0ea-6fd150a6b694","Type":"ContainerDied","Data":"1e2a8215bdb935ebf2c664cc0dd23a1c7f125be5044a587bb13298fd9deeeb59"} Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.980746 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e2a8215bdb935ebf2c664cc0dd23a1c7f125be5044a587bb13298fd9deeeb59" Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.983162 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f761-account-create-update-dmmw8" event={"ID":"87188566-f3cf-4cea-bdd1-28f3c48f5b37","Type":"ContainerDied","Data":"b71d42a3470b120b418e53a61227264d1b6747ff7e6250de550a62d0d7b09a21"} Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.983186 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b71d42a3470b120b418e53a61227264d1b6747ff7e6250de550a62d0d7b09a21" Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.998116 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6nxpn" event={"ID":"0a398e94-3f7a-48ec-815f-79c5a095c9c6","Type":"ContainerDied","Data":"7bfde055b414be05268fc89772325f6880c62de48bf86f4af3c1297f2db3c33d"} Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.998369 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bfde055b414be05268fc89772325f6880c62de48bf86f4af3c1297f2db3c33d" Nov 24 17:44:34 crc kubenswrapper[4808]: I1124 17:44:34.998586 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6nxpn" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.036525 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-dvnqp" event={"ID":"5b835d11-9481-4841-8a77-cc923fbb999b","Type":"ContainerDied","Data":"e194de5e57855ecaf0a3f745c556aa64784576d24d1afb02205f8bb0d24a7d58"} Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.036567 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e194de5e57855ecaf0a3f745c556aa64784576d24d1afb02205f8bb0d24a7d58" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.036665 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-dvnqp" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.036877 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4a1f-account-create-update-tp9xc" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.050194 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17d53839-6156-4d7e-ad19-b8a092398cf6","Type":"ContainerStarted","Data":"77e3535d1a6c69ccfe6e21c36cabf67f3aeebbb0bf6b464fb4a646b3a5c982f3"} Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.050460 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="ceilometer-central-agent" containerID="cri-o://f4141ea6636038d9a14bd6b74666b2906d67f622759fc2868403445ac565f3d3" gracePeriod=30 Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.050815 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.050914 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="proxy-httpd" containerID="cri-o://77e3535d1a6c69ccfe6e21c36cabf67f3aeebbb0bf6b464fb4a646b3a5c982f3" gracePeriod=30 Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.050997 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="sg-core" containerID="cri-o://479c76abd79606ada4f9033e0dd9ff1fdffd1637e35428609952b3ddb85d3b60" gracePeriod=30 Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.051076 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="ceilometer-notification-agent" containerID="cri-o://ae8ce142e42eab69b11061be4c8174ce4dc897d41fb87e7aa1e6317952b8d09a" gracePeriod=30 Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.064299 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b835d11-9481-4841-8a77-cc923fbb999b-operator-scripts\") pod \"5b835d11-9481-4841-8a77-cc923fbb999b\" (UID: \"5b835d11-9481-4841-8a77-cc923fbb999b\") " Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.064629 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2txw\" (UniqueName: \"kubernetes.io/projected/5b835d11-9481-4841-8a77-cc923fbb999b-kube-api-access-s2txw\") pod \"5b835d11-9481-4841-8a77-cc923fbb999b\" (UID: \"5b835d11-9481-4841-8a77-cc923fbb999b\") " Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.064704 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"88cb8610-eb5e-45de-8028-37fc361beaac","Type":"ContainerStarted","Data":"6ea3ee359cd4c8165c1be84babb6bdaf332e73e9f7a807693dbb739a908d03b5"} Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.066297 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b835d11-9481-4841-8a77-cc923fbb999b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5b835d11-9481-4841-8a77-cc923fbb999b" (UID: "5b835d11-9481-4841-8a77-cc923fbb999b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.085150 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b835d11-9481-4841-8a77-cc923fbb999b-kube-api-access-s2txw" (OuterVolumeSpecName: "kube-api-access-s2txw") pod "5b835d11-9481-4841-8a77-cc923fbb999b" (UID: "5b835d11-9481-4841-8a77-cc923fbb999b"). InnerVolumeSpecName "kube-api-access-s2txw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.086566 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-7czsx" event={"ID":"09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d","Type":"ContainerDied","Data":"47489a74be2e33f1acec36e8fe236864596b87db12d6939e681e2b67d70aa310"} Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.086612 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47489a74be2e33f1acec36e8fe236864596b87db12d6939e681e2b67d70aa310" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.101607 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-cec5-account-create-update-vzfxt" event={"ID":"76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c","Type":"ContainerDied","Data":"8e45f0b2368e00215e50f9ca576e984b6ee9eefb73313a1cb3b15d0e1c465db2"} Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.101652 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e45f0b2368e00215e50f9ca576e984b6ee9eefb73313a1cb3b15d0e1c465db2" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.101739 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-cec5-account-create-update-vzfxt" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.105556 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f761-account-create-update-dmmw8" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.109582 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc25c6d7-a302-4ea0-bd25-2b19f9b17a71","Type":"ContainerStarted","Data":"c89a908af1e0606301f1ea32438a27bc9c3aa4a0be868c20084c899d060ea9ee"} Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.112912 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.318889972 podStartE2EDuration="7.112871769s" podCreationTimestamp="2025-11-24 17:44:28 +0000 UTC" firstStartedPulling="2025-11-24 17:44:29.472088532 +0000 UTC m=+1062.069756334" lastFinishedPulling="2025-11-24 17:44:34.266070329 +0000 UTC m=+1066.863738131" observedRunningTime="2025-11-24 17:44:35.08071223 +0000 UTC m=+1067.678380052" watchObservedRunningTime="2025-11-24 17:44:35.112871769 +0000 UTC m=+1067.710539571" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.153826 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.153809704 podStartE2EDuration="4.153809704s" podCreationTimestamp="2025-11-24 17:44:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:44:35.149817578 +0000 UTC m=+1067.747485380" watchObservedRunningTime="2025-11-24 17:44:35.153809704 +0000 UTC m=+1067.751477506" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.167197 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2txw\" (UniqueName: \"kubernetes.io/projected/5b835d11-9481-4841-8a77-cc923fbb999b-kube-api-access-s2txw\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.167225 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b835d11-9481-4841-8a77-cc923fbb999b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.176259 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7czsx" Nov 24 17:44:35 crc kubenswrapper[4808]: E1124 17:44:35.180700 4808 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a398e94_3f7a_48ec_815f_79c5a095c9c6.slice\": RecentStats: unable to find data in memory cache]" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.268588 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vslv9\" (UniqueName: \"kubernetes.io/projected/752ed614-a980-483f-a0ea-6fd150a6b694-kube-api-access-vslv9\") pod \"752ed614-a980-483f-a0ea-6fd150a6b694\" (UID: \"752ed614-a980-483f-a0ea-6fd150a6b694\") " Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.269351 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87188566-f3cf-4cea-bdd1-28f3c48f5b37-operator-scripts\") pod \"87188566-f3cf-4cea-bdd1-28f3c48f5b37\" (UID: \"87188566-f3cf-4cea-bdd1-28f3c48f5b37\") " Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.269559 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d-operator-scripts\") pod \"09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d\" (UID: \"09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d\") " Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.269774 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szwr2\" (UniqueName: \"kubernetes.io/projected/87188566-f3cf-4cea-bdd1-28f3c48f5b37-kube-api-access-szwr2\") pod \"87188566-f3cf-4cea-bdd1-28f3c48f5b37\" (UID: \"87188566-f3cf-4cea-bdd1-28f3c48f5b37\") " Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.269906 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8xjf\" (UniqueName: \"kubernetes.io/projected/09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d-kube-api-access-c8xjf\") pod \"09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d\" (UID: \"09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d\") " Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.270209 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/752ed614-a980-483f-a0ea-6fd150a6b694-operator-scripts\") pod \"752ed614-a980-483f-a0ea-6fd150a6b694\" (UID: \"752ed614-a980-483f-a0ea-6fd150a6b694\") " Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.270471 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d" (UID: "09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.270648 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87188566-f3cf-4cea-bdd1-28f3c48f5b37-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "87188566-f3cf-4cea-bdd1-28f3c48f5b37" (UID: "87188566-f3cf-4cea-bdd1-28f3c48f5b37"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.271985 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.272525 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87188566-f3cf-4cea-bdd1-28f3c48f5b37-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.272100 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/752ed614-a980-483f-a0ea-6fd150a6b694-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "752ed614-a980-483f-a0ea-6fd150a6b694" (UID: "752ed614-a980-483f-a0ea-6fd150a6b694"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.272620 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/752ed614-a980-483f-a0ea-6fd150a6b694-kube-api-access-vslv9" (OuterVolumeSpecName: "kube-api-access-vslv9") pod "752ed614-a980-483f-a0ea-6fd150a6b694" (UID: "752ed614-a980-483f-a0ea-6fd150a6b694"). InnerVolumeSpecName "kube-api-access-vslv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.273077 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87188566-f3cf-4cea-bdd1-28f3c48f5b37-kube-api-access-szwr2" (OuterVolumeSpecName: "kube-api-access-szwr2") pod "87188566-f3cf-4cea-bdd1-28f3c48f5b37" (UID: "87188566-f3cf-4cea-bdd1-28f3c48f5b37"). InnerVolumeSpecName "kube-api-access-szwr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.275490 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d-kube-api-access-c8xjf" (OuterVolumeSpecName: "kube-api-access-c8xjf") pod "09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d" (UID: "09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d"). InnerVolumeSpecName "kube-api-access-c8xjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.374573 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vslv9\" (UniqueName: \"kubernetes.io/projected/752ed614-a980-483f-a0ea-6fd150a6b694-kube-api-access-vslv9\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.374613 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szwr2\" (UniqueName: \"kubernetes.io/projected/87188566-f3cf-4cea-bdd1-28f3c48f5b37-kube-api-access-szwr2\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.374728 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8xjf\" (UniqueName: \"kubernetes.io/projected/09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d-kube-api-access-c8xjf\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:35 crc kubenswrapper[4808]: I1124 17:44:35.374940 4808 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/752ed614-a980-483f-a0ea-6fd150a6b694-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:36 crc kubenswrapper[4808]: I1124 17:44:36.123297 4808 generic.go:334] "Generic (PLEG): container finished" podID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerID="77e3535d1a6c69ccfe6e21c36cabf67f3aeebbb0bf6b464fb4a646b3a5c982f3" exitCode=0 Nov 24 17:44:36 crc kubenswrapper[4808]: I1124 17:44:36.123587 4808 generic.go:334] "Generic (PLEG): container finished" podID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerID="479c76abd79606ada4f9033e0dd9ff1fdffd1637e35428609952b3ddb85d3b60" exitCode=2 Nov 24 17:44:36 crc kubenswrapper[4808]: I1124 17:44:36.123597 4808 generic.go:334] "Generic (PLEG): container finished" podID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerID="ae8ce142e42eab69b11061be4c8174ce4dc897d41fb87e7aa1e6317952b8d09a" exitCode=0 Nov 24 17:44:36 crc kubenswrapper[4808]: I1124 17:44:36.123338 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17d53839-6156-4d7e-ad19-b8a092398cf6","Type":"ContainerDied","Data":"77e3535d1a6c69ccfe6e21c36cabf67f3aeebbb0bf6b464fb4a646b3a5c982f3"} Nov 24 17:44:36 crc kubenswrapper[4808]: I1124 17:44:36.123672 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17d53839-6156-4d7e-ad19-b8a092398cf6","Type":"ContainerDied","Data":"479c76abd79606ada4f9033e0dd9ff1fdffd1637e35428609952b3ddb85d3b60"} Nov 24 17:44:36 crc kubenswrapper[4808]: I1124 17:44:36.123688 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17d53839-6156-4d7e-ad19-b8a092398cf6","Type":"ContainerDied","Data":"ae8ce142e42eab69b11061be4c8174ce4dc897d41fb87e7aa1e6317952b8d09a"} Nov 24 17:44:36 crc kubenswrapper[4808]: I1124 17:44:36.125977 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7czsx" Nov 24 17:44:36 crc kubenswrapper[4808]: I1124 17:44:36.126057 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"88cb8610-eb5e-45de-8028-37fc361beaac","Type":"ContainerStarted","Data":"756052ffc4e6c97376ac859895a1224cad2f5719c1b8bceeb3dc76753d66a5e3"} Nov 24 17:44:36 crc kubenswrapper[4808]: I1124 17:44:36.126124 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"88cb8610-eb5e-45de-8028-37fc361beaac","Type":"ContainerStarted","Data":"79ef0d487930067497a5e1313122a2505de4ccc3c0333366f43626ae96649a20"} Nov 24 17:44:36 crc kubenswrapper[4808]: I1124 17:44:36.126154 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f761-account-create-update-dmmw8" Nov 24 17:44:36 crc kubenswrapper[4808]: I1124 17:44:36.126269 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4a1f-account-create-update-tp9xc" Nov 24 17:44:36 crc kubenswrapper[4808]: I1124 17:44:36.169551 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.169530095 podStartE2EDuration="3.169530095s" podCreationTimestamp="2025-11-24 17:44:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:44:36.155842935 +0000 UTC m=+1068.753510757" watchObservedRunningTime="2025-11-24 17:44:36.169530095 +0000 UTC m=+1068.767197897" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.007544 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.122224 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.166752 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-config-data\") pod \"17d53839-6156-4d7e-ad19-b8a092398cf6\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.166834 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-scripts\") pod \"17d53839-6156-4d7e-ad19-b8a092398cf6\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.166883 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17d53839-6156-4d7e-ad19-b8a092398cf6-log-httpd\") pod \"17d53839-6156-4d7e-ad19-b8a092398cf6\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.166911 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17d53839-6156-4d7e-ad19-b8a092398cf6-run-httpd\") pod \"17d53839-6156-4d7e-ad19-b8a092398cf6\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.166979 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-sg-core-conf-yaml\") pod \"17d53839-6156-4d7e-ad19-b8a092398cf6\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.167073 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwwg2\" (UniqueName: \"kubernetes.io/projected/17d53839-6156-4d7e-ad19-b8a092398cf6-kube-api-access-mwwg2\") pod \"17d53839-6156-4d7e-ad19-b8a092398cf6\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.167143 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-combined-ca-bundle\") pod \"17d53839-6156-4d7e-ad19-b8a092398cf6\" (UID: \"17d53839-6156-4d7e-ad19-b8a092398cf6\") " Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.169262 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17d53839-6156-4d7e-ad19-b8a092398cf6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "17d53839-6156-4d7e-ad19-b8a092398cf6" (UID: "17d53839-6156-4d7e-ad19-b8a092398cf6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.171920 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17d53839-6156-4d7e-ad19-b8a092398cf6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "17d53839-6156-4d7e-ad19-b8a092398cf6" (UID: "17d53839-6156-4d7e-ad19-b8a092398cf6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.173181 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-scripts" (OuterVolumeSpecName: "scripts") pod "17d53839-6156-4d7e-ad19-b8a092398cf6" (UID: "17d53839-6156-4d7e-ad19-b8a092398cf6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.175592 4808 generic.go:334] "Generic (PLEG): container finished" podID="da69b7d2-5a35-4dd7-accb-937fefd0d94f" containerID="cf560a8ffc4af2db3658fa2946d980a7f4fc1a2402bb6f325a229da64eb2fad6" exitCode=137 Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.175669 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74bc8cd64-b5zpt" event={"ID":"da69b7d2-5a35-4dd7-accb-937fefd0d94f","Type":"ContainerDied","Data":"cf560a8ffc4af2db3658fa2946d980a7f4fc1a2402bb6f325a229da64eb2fad6"} Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.175698 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74bc8cd64-b5zpt" event={"ID":"da69b7d2-5a35-4dd7-accb-937fefd0d94f","Type":"ContainerDied","Data":"9ae975cae99004f67fe3732de7bf04f15dab40f283df331982a18682869c14a9"} Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.175714 4808 scope.go:117] "RemoveContainer" containerID="e9322e86325f42914f21f1bb36ca087701f02f1c1f709ab3cf501aaab652f98f" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.175879 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74bc8cd64-b5zpt" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.176732 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17d53839-6156-4d7e-ad19-b8a092398cf6-kube-api-access-mwwg2" (OuterVolumeSpecName: "kube-api-access-mwwg2") pod "17d53839-6156-4d7e-ad19-b8a092398cf6" (UID: "17d53839-6156-4d7e-ad19-b8a092398cf6"). InnerVolumeSpecName "kube-api-access-mwwg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.188147 4808 generic.go:334] "Generic (PLEG): container finished" podID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerID="f4141ea6636038d9a14bd6b74666b2906d67f622759fc2868403445ac565f3d3" exitCode=0 Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.188206 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17d53839-6156-4d7e-ad19-b8a092398cf6","Type":"ContainerDied","Data":"f4141ea6636038d9a14bd6b74666b2906d67f622759fc2868403445ac565f3d3"} Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.188237 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17d53839-6156-4d7e-ad19-b8a092398cf6","Type":"ContainerDied","Data":"5e47702aa44cc848afe866721afb95ff31456a8c2f74d9d9041a15ced65976be"} Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.188308 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.198873 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "17d53839-6156-4d7e-ad19-b8a092398cf6" (UID: "17d53839-6156-4d7e-ad19-b8a092398cf6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.241765 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17d53839-6156-4d7e-ad19-b8a092398cf6" (UID: "17d53839-6156-4d7e-ad19-b8a092398cf6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.268477 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-horizon-tls-certs\") pod \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.268594 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-combined-ca-bundle\") pod \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.268621 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da69b7d2-5a35-4dd7-accb-937fefd0d94f-logs\") pod \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.268695 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-horizon-secret-key\") pod \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.268749 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da69b7d2-5a35-4dd7-accb-937fefd0d94f-scripts\") pod \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.268794 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pbbs\" (UniqueName: \"kubernetes.io/projected/da69b7d2-5a35-4dd7-accb-937fefd0d94f-kube-api-access-7pbbs\") pod \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.268810 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da69b7d2-5a35-4dd7-accb-937fefd0d94f-config-data\") pod \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\" (UID: \"da69b7d2-5a35-4dd7-accb-937fefd0d94f\") " Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.269221 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.269241 4808 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17d53839-6156-4d7e-ad19-b8a092398cf6-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.269252 4808 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17d53839-6156-4d7e-ad19-b8a092398cf6-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.269263 4808 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.269276 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwwg2\" (UniqueName: \"kubernetes.io/projected/17d53839-6156-4d7e-ad19-b8a092398cf6-kube-api-access-mwwg2\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.269286 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.269609 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-config-data" (OuterVolumeSpecName: "config-data") pod "17d53839-6156-4d7e-ad19-b8a092398cf6" (UID: "17d53839-6156-4d7e-ad19-b8a092398cf6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.269662 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da69b7d2-5a35-4dd7-accb-937fefd0d94f-logs" (OuterVolumeSpecName: "logs") pod "da69b7d2-5a35-4dd7-accb-937fefd0d94f" (UID: "da69b7d2-5a35-4dd7-accb-937fefd0d94f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.272024 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da69b7d2-5a35-4dd7-accb-937fefd0d94f-kube-api-access-7pbbs" (OuterVolumeSpecName: "kube-api-access-7pbbs") pod "da69b7d2-5a35-4dd7-accb-937fefd0d94f" (UID: "da69b7d2-5a35-4dd7-accb-937fefd0d94f"). InnerVolumeSpecName "kube-api-access-7pbbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.273216 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "da69b7d2-5a35-4dd7-accb-937fefd0d94f" (UID: "da69b7d2-5a35-4dd7-accb-937fefd0d94f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.294140 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da69b7d2-5a35-4dd7-accb-937fefd0d94f-config-data" (OuterVolumeSpecName: "config-data") pod "da69b7d2-5a35-4dd7-accb-937fefd0d94f" (UID: "da69b7d2-5a35-4dd7-accb-937fefd0d94f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.294901 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da69b7d2-5a35-4dd7-accb-937fefd0d94f-scripts" (OuterVolumeSpecName: "scripts") pod "da69b7d2-5a35-4dd7-accb-937fefd0d94f" (UID: "da69b7d2-5a35-4dd7-accb-937fefd0d94f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.297298 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da69b7d2-5a35-4dd7-accb-937fefd0d94f" (UID: "da69b7d2-5a35-4dd7-accb-937fefd0d94f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.320062 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "da69b7d2-5a35-4dd7-accb-937fefd0d94f" (UID: "da69b7d2-5a35-4dd7-accb-937fefd0d94f"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.371463 4808 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.371507 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17d53839-6156-4d7e-ad19-b8a092398cf6-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.371517 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.371531 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da69b7d2-5a35-4dd7-accb-937fefd0d94f-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.371541 4808 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/da69b7d2-5a35-4dd7-accb-937fefd0d94f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.371551 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da69b7d2-5a35-4dd7-accb-937fefd0d94f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.371562 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pbbs\" (UniqueName: \"kubernetes.io/projected/da69b7d2-5a35-4dd7-accb-937fefd0d94f-kube-api-access-7pbbs\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.371574 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da69b7d2-5a35-4dd7-accb-937fefd0d94f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.372268 4808 scope.go:117] "RemoveContainer" containerID="cf560a8ffc4af2db3658fa2946d980a7f4fc1a2402bb6f325a229da64eb2fad6" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.395688 4808 scope.go:117] "RemoveContainer" containerID="e9322e86325f42914f21f1bb36ca087701f02f1c1f709ab3cf501aaab652f98f" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.396096 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9322e86325f42914f21f1bb36ca087701f02f1c1f709ab3cf501aaab652f98f\": container with ID starting with e9322e86325f42914f21f1bb36ca087701f02f1c1f709ab3cf501aaab652f98f not found: ID does not exist" containerID="e9322e86325f42914f21f1bb36ca087701f02f1c1f709ab3cf501aaab652f98f" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.396150 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9322e86325f42914f21f1bb36ca087701f02f1c1f709ab3cf501aaab652f98f"} err="failed to get container status \"e9322e86325f42914f21f1bb36ca087701f02f1c1f709ab3cf501aaab652f98f\": rpc error: code = NotFound desc = could not find container \"e9322e86325f42914f21f1bb36ca087701f02f1c1f709ab3cf501aaab652f98f\": container with ID starting with e9322e86325f42914f21f1bb36ca087701f02f1c1f709ab3cf501aaab652f98f not found: ID does not exist" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.396175 4808 scope.go:117] "RemoveContainer" containerID="cf560a8ffc4af2db3658fa2946d980a7f4fc1a2402bb6f325a229da64eb2fad6" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.396526 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf560a8ffc4af2db3658fa2946d980a7f4fc1a2402bb6f325a229da64eb2fad6\": container with ID starting with cf560a8ffc4af2db3658fa2946d980a7f4fc1a2402bb6f325a229da64eb2fad6 not found: ID does not exist" containerID="cf560a8ffc4af2db3658fa2946d980a7f4fc1a2402bb6f325a229da64eb2fad6" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.396550 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf560a8ffc4af2db3658fa2946d980a7f4fc1a2402bb6f325a229da64eb2fad6"} err="failed to get container status \"cf560a8ffc4af2db3658fa2946d980a7f4fc1a2402bb6f325a229da64eb2fad6\": rpc error: code = NotFound desc = could not find container \"cf560a8ffc4af2db3658fa2946d980a7f4fc1a2402bb6f325a229da64eb2fad6\": container with ID starting with cf560a8ffc4af2db3658fa2946d980a7f4fc1a2402bb6f325a229da64eb2fad6 not found: ID does not exist" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.396564 4808 scope.go:117] "RemoveContainer" containerID="77e3535d1a6c69ccfe6e21c36cabf67f3aeebbb0bf6b464fb4a646b3a5c982f3" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.412138 4808 scope.go:117] "RemoveContainer" containerID="479c76abd79606ada4f9033e0dd9ff1fdffd1637e35428609952b3ddb85d3b60" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.429751 4808 scope.go:117] "RemoveContainer" containerID="ae8ce142e42eab69b11061be4c8174ce4dc897d41fb87e7aa1e6317952b8d09a" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.448983 4808 scope.go:117] "RemoveContainer" containerID="f4141ea6636038d9a14bd6b74666b2906d67f622759fc2868403445ac565f3d3" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.504326 4808 scope.go:117] "RemoveContainer" containerID="77e3535d1a6c69ccfe6e21c36cabf67f3aeebbb0bf6b464fb4a646b3a5c982f3" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.504926 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77e3535d1a6c69ccfe6e21c36cabf67f3aeebbb0bf6b464fb4a646b3a5c982f3\": container with ID starting with 77e3535d1a6c69ccfe6e21c36cabf67f3aeebbb0bf6b464fb4a646b3a5c982f3 not found: ID does not exist" containerID="77e3535d1a6c69ccfe6e21c36cabf67f3aeebbb0bf6b464fb4a646b3a5c982f3" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.504972 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77e3535d1a6c69ccfe6e21c36cabf67f3aeebbb0bf6b464fb4a646b3a5c982f3"} err="failed to get container status \"77e3535d1a6c69ccfe6e21c36cabf67f3aeebbb0bf6b464fb4a646b3a5c982f3\": rpc error: code = NotFound desc = could not find container \"77e3535d1a6c69ccfe6e21c36cabf67f3aeebbb0bf6b464fb4a646b3a5c982f3\": container with ID starting with 77e3535d1a6c69ccfe6e21c36cabf67f3aeebbb0bf6b464fb4a646b3a5c982f3 not found: ID does not exist" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.505004 4808 scope.go:117] "RemoveContainer" containerID="479c76abd79606ada4f9033e0dd9ff1fdffd1637e35428609952b3ddb85d3b60" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.505500 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"479c76abd79606ada4f9033e0dd9ff1fdffd1637e35428609952b3ddb85d3b60\": container with ID starting with 479c76abd79606ada4f9033e0dd9ff1fdffd1637e35428609952b3ddb85d3b60 not found: ID does not exist" containerID="479c76abd79606ada4f9033e0dd9ff1fdffd1637e35428609952b3ddb85d3b60" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.505550 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"479c76abd79606ada4f9033e0dd9ff1fdffd1637e35428609952b3ddb85d3b60"} err="failed to get container status \"479c76abd79606ada4f9033e0dd9ff1fdffd1637e35428609952b3ddb85d3b60\": rpc error: code = NotFound desc = could not find container \"479c76abd79606ada4f9033e0dd9ff1fdffd1637e35428609952b3ddb85d3b60\": container with ID starting with 479c76abd79606ada4f9033e0dd9ff1fdffd1637e35428609952b3ddb85d3b60 not found: ID does not exist" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.505582 4808 scope.go:117] "RemoveContainer" containerID="ae8ce142e42eab69b11061be4c8174ce4dc897d41fb87e7aa1e6317952b8d09a" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.506438 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74bc8cd64-b5zpt"] Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.506796 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae8ce142e42eab69b11061be4c8174ce4dc897d41fb87e7aa1e6317952b8d09a\": container with ID starting with ae8ce142e42eab69b11061be4c8174ce4dc897d41fb87e7aa1e6317952b8d09a not found: ID does not exist" containerID="ae8ce142e42eab69b11061be4c8174ce4dc897d41fb87e7aa1e6317952b8d09a" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.506828 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae8ce142e42eab69b11061be4c8174ce4dc897d41fb87e7aa1e6317952b8d09a"} err="failed to get container status \"ae8ce142e42eab69b11061be4c8174ce4dc897d41fb87e7aa1e6317952b8d09a\": rpc error: code = NotFound desc = could not find container \"ae8ce142e42eab69b11061be4c8174ce4dc897d41fb87e7aa1e6317952b8d09a\": container with ID starting with ae8ce142e42eab69b11061be4c8174ce4dc897d41fb87e7aa1e6317952b8d09a not found: ID does not exist" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.506848 4808 scope.go:117] "RemoveContainer" containerID="f4141ea6636038d9a14bd6b74666b2906d67f622759fc2868403445ac565f3d3" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.507495 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4141ea6636038d9a14bd6b74666b2906d67f622759fc2868403445ac565f3d3\": container with ID starting with f4141ea6636038d9a14bd6b74666b2906d67f622759fc2868403445ac565f3d3 not found: ID does not exist" containerID="f4141ea6636038d9a14bd6b74666b2906d67f622759fc2868403445ac565f3d3" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.507542 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4141ea6636038d9a14bd6b74666b2906d67f622759fc2868403445ac565f3d3"} err="failed to get container status \"f4141ea6636038d9a14bd6b74666b2906d67f622759fc2868403445ac565f3d3\": rpc error: code = NotFound desc = could not find container \"f4141ea6636038d9a14bd6b74666b2906d67f622759fc2868403445ac565f3d3\": container with ID starting with f4141ea6636038d9a14bd6b74666b2906d67f622759fc2868403445ac565f3d3 not found: ID does not exist" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.520995 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-74bc8cd64-b5zpt"] Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.528614 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.550313 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.562908 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.563357 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da69b7d2-5a35-4dd7-accb-937fefd0d94f" containerName="horizon" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563374 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="da69b7d2-5a35-4dd7-accb-937fefd0d94f" containerName="horizon" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.563382 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="ceilometer-notification-agent" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563388 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="ceilometer-notification-agent" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.563408 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87188566-f3cf-4cea-bdd1-28f3c48f5b37" containerName="mariadb-account-create-update" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563414 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="87188566-f3cf-4cea-bdd1-28f3c48f5b37" containerName="mariadb-account-create-update" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.563421 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d" containerName="mariadb-database-create" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563427 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d" containerName="mariadb-database-create" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.563438 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da69b7d2-5a35-4dd7-accb-937fefd0d94f" containerName="horizon-log" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563444 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="da69b7d2-5a35-4dd7-accb-937fefd0d94f" containerName="horizon-log" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.563452 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="sg-core" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563458 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="sg-core" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.563473 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="752ed614-a980-483f-a0ea-6fd150a6b694" containerName="mariadb-account-create-update" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563480 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="752ed614-a980-483f-a0ea-6fd150a6b694" containerName="mariadb-account-create-update" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.563495 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a398e94-3f7a-48ec-815f-79c5a095c9c6" containerName="mariadb-database-create" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563503 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a398e94-3f7a-48ec-815f-79c5a095c9c6" containerName="mariadb-database-create" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.563513 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b835d11-9481-4841-8a77-cc923fbb999b" containerName="mariadb-database-create" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563518 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b835d11-9481-4841-8a77-cc923fbb999b" containerName="mariadb-database-create" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.563532 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="proxy-httpd" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563538 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="proxy-httpd" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.563554 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="ceilometer-central-agent" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563559 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="ceilometer-central-agent" Nov 24 17:44:40 crc kubenswrapper[4808]: E1124 17:44:40.563567 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c" containerName="mariadb-account-create-update" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563572 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c" containerName="mariadb-account-create-update" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563727 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d" containerName="mariadb-database-create" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563740 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="sg-core" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563754 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="ceilometer-notification-agent" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563762 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="ceilometer-central-agent" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563773 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="da69b7d2-5a35-4dd7-accb-937fefd0d94f" containerName="horizon" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563782 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="da69b7d2-5a35-4dd7-accb-937fefd0d94f" containerName="horizon-log" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563792 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="752ed614-a980-483f-a0ea-6fd150a6b694" containerName="mariadb-account-create-update" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563799 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a398e94-3f7a-48ec-815f-79c5a095c9c6" containerName="mariadb-database-create" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563806 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c" containerName="mariadb-account-create-update" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563815 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" containerName="proxy-httpd" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563825 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="87188566-f3cf-4cea-bdd1-28f3c48f5b37" containerName="mariadb-account-create-update" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.563835 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b835d11-9481-4841-8a77-cc923fbb999b" containerName="mariadb-database-create" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.565510 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.571629 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.571774 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.575801 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.676697 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5sjm\" (UniqueName: \"kubernetes.io/projected/7ebea6b3-5803-49a8-a846-24e716c7f772-kube-api-access-f5sjm\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.676738 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.676805 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ebea6b3-5803-49a8-a846-24e716c7f772-run-httpd\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.676825 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-scripts\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.676868 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-config-data\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.676908 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.676926 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ebea6b3-5803-49a8-a846-24e716c7f772-log-httpd\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.778577 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5sjm\" (UniqueName: \"kubernetes.io/projected/7ebea6b3-5803-49a8-a846-24e716c7f772-kube-api-access-f5sjm\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.778630 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.778699 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ebea6b3-5803-49a8-a846-24e716c7f772-run-httpd\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.778720 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-scripts\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.778780 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-config-data\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.778836 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.778863 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ebea6b3-5803-49a8-a846-24e716c7f772-log-httpd\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.779311 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ebea6b3-5803-49a8-a846-24e716c7f772-log-httpd\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.781173 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ebea6b3-5803-49a8-a846-24e716c7f772-run-httpd\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.785144 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.785215 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-scripts\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.785589 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.798141 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-config-data\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.811643 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5sjm\" (UniqueName: \"kubernetes.io/projected/7ebea6b3-5803-49a8-a846-24e716c7f772-kube-api-access-f5sjm\") pod \"ceilometer-0\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.900994 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.911791 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xz9xq"] Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.913372 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.916914 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.916959 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.917232 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-b5s49" Nov 24 17:44:40 crc kubenswrapper[4808]: I1124 17:44:40.926098 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xz9xq"] Nov 24 17:44:41 crc kubenswrapper[4808]: I1124 17:44:41.084729 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-scripts\") pod \"nova-cell0-conductor-db-sync-xz9xq\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:44:41 crc kubenswrapper[4808]: I1124 17:44:41.085096 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xz9xq\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:44:41 crc kubenswrapper[4808]: I1124 17:44:41.085306 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-config-data\") pod \"nova-cell0-conductor-db-sync-xz9xq\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:44:41 crc kubenswrapper[4808]: I1124 17:44:41.085344 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9gtk\" (UniqueName: \"kubernetes.io/projected/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-kube-api-access-z9gtk\") pod \"nova-cell0-conductor-db-sync-xz9xq\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:44:41 crc kubenswrapper[4808]: I1124 17:44:41.187995 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-config-data\") pod \"nova-cell0-conductor-db-sync-xz9xq\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:44:41 crc kubenswrapper[4808]: I1124 17:44:41.188134 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9gtk\" (UniqueName: \"kubernetes.io/projected/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-kube-api-access-z9gtk\") pod \"nova-cell0-conductor-db-sync-xz9xq\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:44:41 crc kubenswrapper[4808]: I1124 17:44:41.188232 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-scripts\") pod \"nova-cell0-conductor-db-sync-xz9xq\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:44:41 crc kubenswrapper[4808]: I1124 17:44:41.188318 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xz9xq\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:44:41 crc kubenswrapper[4808]: I1124 17:44:41.193595 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-scripts\") pod \"nova-cell0-conductor-db-sync-xz9xq\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:44:41 crc kubenswrapper[4808]: I1124 17:44:41.194487 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-config-data\") pod \"nova-cell0-conductor-db-sync-xz9xq\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:44:41 crc kubenswrapper[4808]: I1124 17:44:41.203663 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xz9xq\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:44:41 crc kubenswrapper[4808]: I1124 17:44:41.205690 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9gtk\" (UniqueName: \"kubernetes.io/projected/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-kube-api-access-z9gtk\") pod \"nova-cell0-conductor-db-sync-xz9xq\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:44:41 crc kubenswrapper[4808]: I1124 17:44:41.325125 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:44:41 crc kubenswrapper[4808]: I1124 17:44:41.378294 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:41 crc kubenswrapper[4808]: I1124 17:44:41.798105 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xz9xq"] Nov 24 17:44:42 crc kubenswrapper[4808]: I1124 17:44:42.210430 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 17:44:42 crc kubenswrapper[4808]: I1124 17:44:42.210731 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 17:44:42 crc kubenswrapper[4808]: I1124 17:44:42.222492 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xz9xq" event={"ID":"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510","Type":"ContainerStarted","Data":"d6c6b7fb8076934f76a611688933217bc8881288ad55eabdcb622927d196fb26"} Nov 24 17:44:42 crc kubenswrapper[4808]: I1124 17:44:42.224695 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ebea6b3-5803-49a8-a846-24e716c7f772","Type":"ContainerStarted","Data":"f4de3c099dd104064863cc7acb3ee60247d9a0f6e6b664a88588bce93031fbac"} Nov 24 17:44:42 crc kubenswrapper[4808]: I1124 17:44:42.224731 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ebea6b3-5803-49a8-a846-24e716c7f772","Type":"ContainerStarted","Data":"b33f0ce371b2cfd3acf975cf468137b9e78d2b020bab7551a654bff435684748"} Nov 24 17:44:42 crc kubenswrapper[4808]: I1124 17:44:42.240857 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 17:44:42 crc kubenswrapper[4808]: I1124 17:44:42.253604 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 17:44:42 crc kubenswrapper[4808]: I1124 17:44:42.357327 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17d53839-6156-4d7e-ad19-b8a092398cf6" path="/var/lib/kubelet/pods/17d53839-6156-4d7e-ad19-b8a092398cf6/volumes" Nov 24 17:44:42 crc kubenswrapper[4808]: I1124 17:44:42.358235 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da69b7d2-5a35-4dd7-accb-937fefd0d94f" path="/var/lib/kubelet/pods/da69b7d2-5a35-4dd7-accb-937fefd0d94f/volumes" Nov 24 17:44:43 crc kubenswrapper[4808]: I1124 17:44:43.244789 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ebea6b3-5803-49a8-a846-24e716c7f772","Type":"ContainerStarted","Data":"d1311b450a41ea6c1cc05cd098e9b152301408fef406d455324d36722ab859c6"} Nov 24 17:44:43 crc kubenswrapper[4808]: I1124 17:44:43.245152 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 17:44:43 crc kubenswrapper[4808]: I1124 17:44:43.245302 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 17:44:43 crc kubenswrapper[4808]: I1124 17:44:43.511852 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 17:44:43 crc kubenswrapper[4808]: I1124 17:44:43.512206 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 17:44:43 crc kubenswrapper[4808]: I1124 17:44:43.546281 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 17:44:43 crc kubenswrapper[4808]: I1124 17:44:43.555927 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 17:44:44 crc kubenswrapper[4808]: I1124 17:44:44.260953 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ebea6b3-5803-49a8-a846-24e716c7f772","Type":"ContainerStarted","Data":"36164a982939f3d6fe0ebc8113370d0b7f9d20eb9ef7bc12885d931b9959c9f1"} Nov 24 17:44:44 crc kubenswrapper[4808]: I1124 17:44:44.260994 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 17:44:44 crc kubenswrapper[4808]: I1124 17:44:44.261215 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 17:44:44 crc kubenswrapper[4808]: I1124 17:44:44.690460 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:45 crc kubenswrapper[4808]: I1124 17:44:45.294385 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ebea6b3-5803-49a8-a846-24e716c7f772","Type":"ContainerStarted","Data":"eb5419adca387d56aae336439ace336fee69b28bc2309cce9acbe4d9264f79f0"} Nov 24 17:44:45 crc kubenswrapper[4808]: I1124 17:44:45.294992 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="ceilometer-central-agent" containerID="cri-o://f4de3c099dd104064863cc7acb3ee60247d9a0f6e6b664a88588bce93031fbac" gracePeriod=30 Nov 24 17:44:45 crc kubenswrapper[4808]: I1124 17:44:45.295172 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 17:44:45 crc kubenswrapper[4808]: I1124 17:44:45.295202 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="proxy-httpd" containerID="cri-o://eb5419adca387d56aae336439ace336fee69b28bc2309cce9acbe4d9264f79f0" gracePeriod=30 Nov 24 17:44:45 crc kubenswrapper[4808]: I1124 17:44:45.295260 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="sg-core" containerID="cri-o://36164a982939f3d6fe0ebc8113370d0b7f9d20eb9ef7bc12885d931b9959c9f1" gracePeriod=30 Nov 24 17:44:45 crc kubenswrapper[4808]: I1124 17:44:45.295303 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="ceilometer-notification-agent" containerID="cri-o://d1311b450a41ea6c1cc05cd098e9b152301408fef406d455324d36722ab859c6" gracePeriod=30 Nov 24 17:44:45 crc kubenswrapper[4808]: I1124 17:44:45.318364 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.225668747 podStartE2EDuration="5.318345189s" podCreationTimestamp="2025-11-24 17:44:40 +0000 UTC" firstStartedPulling="2025-11-24 17:44:41.385878632 +0000 UTC m=+1073.983546434" lastFinishedPulling="2025-11-24 17:44:44.478555074 +0000 UTC m=+1077.076222876" observedRunningTime="2025-11-24 17:44:45.31804169 +0000 UTC m=+1077.915709512" watchObservedRunningTime="2025-11-24 17:44:45.318345189 +0000 UTC m=+1077.916012991" Nov 24 17:44:45 crc kubenswrapper[4808]: E1124 17:44:45.444133 4808 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ebea6b3_5803_49a8_a846_24e716c7f772.slice/crio-conmon-36164a982939f3d6fe0ebc8113370d0b7f9d20eb9ef7bc12885d931b9959c9f1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ebea6b3_5803_49a8_a846_24e716c7f772.slice/crio-36164a982939f3d6fe0ebc8113370d0b7f9d20eb9ef7bc12885d931b9959c9f1.scope\": RecentStats: unable to find data in memory cache]" Nov 24 17:44:45 crc kubenswrapper[4808]: I1124 17:44:45.467760 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 17:44:45 crc kubenswrapper[4808]: I1124 17:44:45.468123 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 17:44:46 crc kubenswrapper[4808]: I1124 17:44:46.309693 4808 generic.go:334] "Generic (PLEG): container finished" podID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerID="eb5419adca387d56aae336439ace336fee69b28bc2309cce9acbe4d9264f79f0" exitCode=0 Nov 24 17:44:46 crc kubenswrapper[4808]: I1124 17:44:46.309985 4808 generic.go:334] "Generic (PLEG): container finished" podID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerID="36164a982939f3d6fe0ebc8113370d0b7f9d20eb9ef7bc12885d931b9959c9f1" exitCode=2 Nov 24 17:44:46 crc kubenswrapper[4808]: I1124 17:44:46.309993 4808 generic.go:334] "Generic (PLEG): container finished" podID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerID="d1311b450a41ea6c1cc05cd098e9b152301408fef406d455324d36722ab859c6" exitCode=0 Nov 24 17:44:46 crc kubenswrapper[4808]: I1124 17:44:46.311219 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ebea6b3-5803-49a8-a846-24e716c7f772","Type":"ContainerDied","Data":"eb5419adca387d56aae336439ace336fee69b28bc2309cce9acbe4d9264f79f0"} Nov 24 17:44:46 crc kubenswrapper[4808]: I1124 17:44:46.311251 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ebea6b3-5803-49a8-a846-24e716c7f772","Type":"ContainerDied","Data":"36164a982939f3d6fe0ebc8113370d0b7f9d20eb9ef7bc12885d931b9959c9f1"} Nov 24 17:44:46 crc kubenswrapper[4808]: I1124 17:44:46.311262 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ebea6b3-5803-49a8-a846-24e716c7f772","Type":"ContainerDied","Data":"d1311b450a41ea6c1cc05cd098e9b152301408fef406d455324d36722ab859c6"} Nov 24 17:44:46 crc kubenswrapper[4808]: I1124 17:44:46.311306 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 17:44:46 crc kubenswrapper[4808]: I1124 17:44:46.311315 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 17:44:46 crc kubenswrapper[4808]: I1124 17:44:46.527492 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 17:44:46 crc kubenswrapper[4808]: I1124 17:44:46.528615 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 17:44:51 crc kubenswrapper[4808]: I1124 17:44:51.359486 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xz9xq" event={"ID":"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510","Type":"ContainerStarted","Data":"865fc6bbbfe63477f72e6511cdf09da548421a81f1f2d025c582821f266b8dda"} Nov 24 17:44:51 crc kubenswrapper[4808]: I1124 17:44:51.377129 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-xz9xq" podStartSLOduration=2.461967145 podStartE2EDuration="11.377110897s" podCreationTimestamp="2025-11-24 17:44:40 +0000 UTC" firstStartedPulling="2025-11-24 17:44:41.808745206 +0000 UTC m=+1074.406413008" lastFinishedPulling="2025-11-24 17:44:50.723888958 +0000 UTC m=+1083.321556760" observedRunningTime="2025-11-24 17:44:51.371762681 +0000 UTC m=+1083.969430503" watchObservedRunningTime="2025-11-24 17:44:51.377110897 +0000 UTC m=+1083.974778699" Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.370678 4808 generic.go:334] "Generic (PLEG): container finished" podID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerID="f4de3c099dd104064863cc7acb3ee60247d9a0f6e6b664a88588bce93031fbac" exitCode=0 Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.370706 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ebea6b3-5803-49a8-a846-24e716c7f772","Type":"ContainerDied","Data":"f4de3c099dd104064863cc7acb3ee60247d9a0f6e6b664a88588bce93031fbac"} Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.623641 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.795593 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ebea6b3-5803-49a8-a846-24e716c7f772-run-httpd\") pod \"7ebea6b3-5803-49a8-a846-24e716c7f772\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.795889 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-scripts\") pod \"7ebea6b3-5803-49a8-a846-24e716c7f772\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.795923 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5sjm\" (UniqueName: \"kubernetes.io/projected/7ebea6b3-5803-49a8-a846-24e716c7f772-kube-api-access-f5sjm\") pod \"7ebea6b3-5803-49a8-a846-24e716c7f772\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.795970 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-combined-ca-bundle\") pod \"7ebea6b3-5803-49a8-a846-24e716c7f772\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.796010 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-config-data\") pod \"7ebea6b3-5803-49a8-a846-24e716c7f772\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.796080 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ebea6b3-5803-49a8-a846-24e716c7f772-log-httpd\") pod \"7ebea6b3-5803-49a8-a846-24e716c7f772\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.796108 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-sg-core-conf-yaml\") pod \"7ebea6b3-5803-49a8-a846-24e716c7f772\" (UID: \"7ebea6b3-5803-49a8-a846-24e716c7f772\") " Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.796207 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ebea6b3-5803-49a8-a846-24e716c7f772-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7ebea6b3-5803-49a8-a846-24e716c7f772" (UID: "7ebea6b3-5803-49a8-a846-24e716c7f772"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.796805 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ebea6b3-5803-49a8-a846-24e716c7f772-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7ebea6b3-5803-49a8-a846-24e716c7f772" (UID: "7ebea6b3-5803-49a8-a846-24e716c7f772"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.797008 4808 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ebea6b3-5803-49a8-a846-24e716c7f772-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.797041 4808 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ebea6b3-5803-49a8-a846-24e716c7f772-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.803109 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-scripts" (OuterVolumeSpecName: "scripts") pod "7ebea6b3-5803-49a8-a846-24e716c7f772" (UID: "7ebea6b3-5803-49a8-a846-24e716c7f772"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.805634 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ebea6b3-5803-49a8-a846-24e716c7f772-kube-api-access-f5sjm" (OuterVolumeSpecName: "kube-api-access-f5sjm") pod "7ebea6b3-5803-49a8-a846-24e716c7f772" (UID: "7ebea6b3-5803-49a8-a846-24e716c7f772"). InnerVolumeSpecName "kube-api-access-f5sjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.849315 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7ebea6b3-5803-49a8-a846-24e716c7f772" (UID: "7ebea6b3-5803-49a8-a846-24e716c7f772"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.898947 4808 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.898991 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.899004 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5sjm\" (UniqueName: \"kubernetes.io/projected/7ebea6b3-5803-49a8-a846-24e716c7f772-kube-api-access-f5sjm\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.900164 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ebea6b3-5803-49a8-a846-24e716c7f772" (UID: "7ebea6b3-5803-49a8-a846-24e716c7f772"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:52 crc kubenswrapper[4808]: I1124 17:44:52.950896 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-config-data" (OuterVolumeSpecName: "config-data") pod "7ebea6b3-5803-49a8-a846-24e716c7f772" (UID: "7ebea6b3-5803-49a8-a846-24e716c7f772"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.000738 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.000776 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebea6b3-5803-49a8-a846-24e716c7f772-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.383734 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ebea6b3-5803-49a8-a846-24e716c7f772","Type":"ContainerDied","Data":"b33f0ce371b2cfd3acf975cf468137b9e78d2b020bab7551a654bff435684748"} Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.383797 4808 scope.go:117] "RemoveContainer" containerID="eb5419adca387d56aae336439ace336fee69b28bc2309cce9acbe4d9264f79f0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.383831 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.409448 4808 scope.go:117] "RemoveContainer" containerID="36164a982939f3d6fe0ebc8113370d0b7f9d20eb9ef7bc12885d931b9959c9f1" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.435408 4808 scope.go:117] "RemoveContainer" containerID="d1311b450a41ea6c1cc05cd098e9b152301408fef406d455324d36722ab859c6" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.445420 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.465171 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.473872 4808 scope.go:117] "RemoveContainer" containerID="f4de3c099dd104064863cc7acb3ee60247d9a0f6e6b664a88588bce93031fbac" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.475047 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:53 crc kubenswrapper[4808]: E1124 17:44:53.475841 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="proxy-httpd" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.475915 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="proxy-httpd" Nov 24 17:44:53 crc kubenswrapper[4808]: E1124 17:44:53.475980 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="ceilometer-central-agent" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.476059 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="ceilometer-central-agent" Nov 24 17:44:53 crc kubenswrapper[4808]: E1124 17:44:53.476192 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="ceilometer-notification-agent" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.476257 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="ceilometer-notification-agent" Nov 24 17:44:53 crc kubenswrapper[4808]: E1124 17:44:53.476336 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="sg-core" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.476393 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="sg-core" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.476613 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="proxy-httpd" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.476734 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="ceilometer-notification-agent" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.476803 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="sg-core" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.476875 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" containerName="ceilometer-central-agent" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.478899 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.482985 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.483897 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.484506 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.509085 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68962a01-d931-41b7-9492-a2d7fca41788-run-httpd\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.509140 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-scripts\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.509227 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.509263 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l47x\" (UniqueName: \"kubernetes.io/projected/68962a01-d931-41b7-9492-a2d7fca41788-kube-api-access-2l47x\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.509289 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.509322 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68962a01-d931-41b7-9492-a2d7fca41788-log-httpd\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.509377 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-config-data\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.610423 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68962a01-d931-41b7-9492-a2d7fca41788-run-httpd\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.610485 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-scripts\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.610555 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.610591 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l47x\" (UniqueName: \"kubernetes.io/projected/68962a01-d931-41b7-9492-a2d7fca41788-kube-api-access-2l47x\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.610612 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.610629 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68962a01-d931-41b7-9492-a2d7fca41788-log-httpd\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.610679 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-config-data\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.612208 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68962a01-d931-41b7-9492-a2d7fca41788-log-httpd\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.613534 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68962a01-d931-41b7-9492-a2d7fca41788-run-httpd\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.613852 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.614337 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-config-data\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.616611 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.623924 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-scripts\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.647689 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l47x\" (UniqueName: \"kubernetes.io/projected/68962a01-d931-41b7-9492-a2d7fca41788-kube-api-access-2l47x\") pod \"ceilometer-0\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " pod="openstack/ceilometer-0" Nov 24 17:44:53 crc kubenswrapper[4808]: I1124 17:44:53.807037 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:44:54 crc kubenswrapper[4808]: I1124 17:44:54.244320 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:44:54 crc kubenswrapper[4808]: W1124 17:44:54.254733 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68962a01_d931_41b7_9492_a2d7fca41788.slice/crio-218768293c045feaa7e0eaa19b643dff8fb80eca2f7fe872df0856854e283b40 WatchSource:0}: Error finding container 218768293c045feaa7e0eaa19b643dff8fb80eca2f7fe872df0856854e283b40: Status 404 returned error can't find the container with id 218768293c045feaa7e0eaa19b643dff8fb80eca2f7fe872df0856854e283b40 Nov 24 17:44:54 crc kubenswrapper[4808]: I1124 17:44:54.361229 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ebea6b3-5803-49a8-a846-24e716c7f772" path="/var/lib/kubelet/pods/7ebea6b3-5803-49a8-a846-24e716c7f772/volumes" Nov 24 17:44:54 crc kubenswrapper[4808]: I1124 17:44:54.396935 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68962a01-d931-41b7-9492-a2d7fca41788","Type":"ContainerStarted","Data":"218768293c045feaa7e0eaa19b643dff8fb80eca2f7fe872df0856854e283b40"} Nov 24 17:44:55 crc kubenswrapper[4808]: I1124 17:44:55.408886 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68962a01-d931-41b7-9492-a2d7fca41788","Type":"ContainerStarted","Data":"f0675119251252b15ac0b1e0a2f60a6e1291a0f5f9a91d7e0d02507675bc281e"} Nov 24 17:44:57 crc kubenswrapper[4808]: I1124 17:44:57.431456 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68962a01-d931-41b7-9492-a2d7fca41788","Type":"ContainerStarted","Data":"ad58ed120a79c69bdd7a2a466c479e5149fb3ea0c94b1ba71cef7b9334d65a1b"} Nov 24 17:44:58 crc kubenswrapper[4808]: I1124 17:44:58.457230 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68962a01-d931-41b7-9492-a2d7fca41788","Type":"ContainerStarted","Data":"1a3d71f3643a1858ed2ba7b36ca80237b4ec90b8f96fab11420399c418d5ed46"} Nov 24 17:44:59 crc kubenswrapper[4808]: I1124 17:44:59.469813 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68962a01-d931-41b7-9492-a2d7fca41788","Type":"ContainerStarted","Data":"306ce16c3afa8e3fbcae31675bccc7902c4c9f4ff49f7ccfbfda0170eb0d50ee"} Nov 24 17:44:59 crc kubenswrapper[4808]: I1124 17:44:59.470191 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 17:44:59 crc kubenswrapper[4808]: I1124 17:44:59.491341 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.132870036 podStartE2EDuration="6.491319859s" podCreationTimestamp="2025-11-24 17:44:53 +0000 UTC" firstStartedPulling="2025-11-24 17:44:54.257225904 +0000 UTC m=+1086.854893696" lastFinishedPulling="2025-11-24 17:44:58.615675707 +0000 UTC m=+1091.213343519" observedRunningTime="2025-11-24 17:44:59.486370385 +0000 UTC m=+1092.084038187" watchObservedRunningTime="2025-11-24 17:44:59.491319859 +0000 UTC m=+1092.088987681" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.136720 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp"] Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.138528 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.141154 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.142645 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.153564 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp"] Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.237475 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-config-volume\") pod \"collect-profiles-29400105-xfvbp\" (UID: \"71bc118d-05fa-4b8e-84ff-e8b9ad75de68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.237595 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9qwz\" (UniqueName: \"kubernetes.io/projected/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-kube-api-access-l9qwz\") pod \"collect-profiles-29400105-xfvbp\" (UID: \"71bc118d-05fa-4b8e-84ff-e8b9ad75de68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.237616 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-secret-volume\") pod \"collect-profiles-29400105-xfvbp\" (UID: \"71bc118d-05fa-4b8e-84ff-e8b9ad75de68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.338535 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-config-volume\") pod \"collect-profiles-29400105-xfvbp\" (UID: \"71bc118d-05fa-4b8e-84ff-e8b9ad75de68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.338663 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9qwz\" (UniqueName: \"kubernetes.io/projected/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-kube-api-access-l9qwz\") pod \"collect-profiles-29400105-xfvbp\" (UID: \"71bc118d-05fa-4b8e-84ff-e8b9ad75de68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.338687 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-secret-volume\") pod \"collect-profiles-29400105-xfvbp\" (UID: \"71bc118d-05fa-4b8e-84ff-e8b9ad75de68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.339823 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-config-volume\") pod \"collect-profiles-29400105-xfvbp\" (UID: \"71bc118d-05fa-4b8e-84ff-e8b9ad75de68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.345309 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-secret-volume\") pod \"collect-profiles-29400105-xfvbp\" (UID: \"71bc118d-05fa-4b8e-84ff-e8b9ad75de68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.356326 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9qwz\" (UniqueName: \"kubernetes.io/projected/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-kube-api-access-l9qwz\") pod \"collect-profiles-29400105-xfvbp\" (UID: \"71bc118d-05fa-4b8e-84ff-e8b9ad75de68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.456084 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.583141 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="25fd4604-12b2-412f-8d76-15584feda527" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.154:9292/healthcheck\": dial tcp 10.217.0.154:9292: i/o timeout" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.586640 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="25fd4604-12b2-412f-8d76-15584feda527" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.154:9292/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.850976 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:45:00 crc kubenswrapper[4808]: I1124 17:45:00.937639 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp"] Nov 24 17:45:01 crc kubenswrapper[4808]: I1124 17:45:01.496479 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" event={"ID":"71bc118d-05fa-4b8e-84ff-e8b9ad75de68","Type":"ContainerStarted","Data":"a3737b255ab02d41b6d3f0f2d908df7352943c3f48369d2be026048a147c587b"} Nov 24 17:45:01 crc kubenswrapper[4808]: I1124 17:45:01.496891 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" event={"ID":"71bc118d-05fa-4b8e-84ff-e8b9ad75de68","Type":"ContainerStarted","Data":"d3b99ec44910f20f22dc01f6ff3844bd114c54b01e818e41a39f7a3af09123f7"} Nov 24 17:45:01 crc kubenswrapper[4808]: I1124 17:45:01.496780 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="proxy-httpd" containerID="cri-o://306ce16c3afa8e3fbcae31675bccc7902c4c9f4ff49f7ccfbfda0170eb0d50ee" gracePeriod=30 Nov 24 17:45:01 crc kubenswrapper[4808]: I1124 17:45:01.496801 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="sg-core" containerID="cri-o://1a3d71f3643a1858ed2ba7b36ca80237b4ec90b8f96fab11420399c418d5ed46" gracePeriod=30 Nov 24 17:45:01 crc kubenswrapper[4808]: I1124 17:45:01.496801 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="ceilometer-notification-agent" containerID="cri-o://ad58ed120a79c69bdd7a2a466c479e5149fb3ea0c94b1ba71cef7b9334d65a1b" gracePeriod=30 Nov 24 17:45:01 crc kubenswrapper[4808]: I1124 17:45:01.496728 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="ceilometer-central-agent" containerID="cri-o://f0675119251252b15ac0b1e0a2f60a6e1291a0f5f9a91d7e0d02507675bc281e" gracePeriod=30 Nov 24 17:45:01 crc kubenswrapper[4808]: I1124 17:45:01.521755 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" podStartSLOduration=1.521738971 podStartE2EDuration="1.521738971s" podCreationTimestamp="2025-11-24 17:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:45:01.51896363 +0000 UTC m=+1094.116631442" watchObservedRunningTime="2025-11-24 17:45:01.521738971 +0000 UTC m=+1094.119406773" Nov 24 17:45:02 crc kubenswrapper[4808]: I1124 17:45:02.512385 4808 generic.go:334] "Generic (PLEG): container finished" podID="bd6dfdb4-e6c9-4302-bdf5-cbb81348a510" containerID="865fc6bbbfe63477f72e6511cdf09da548421a81f1f2d025c582821f266b8dda" exitCode=0 Nov 24 17:45:02 crc kubenswrapper[4808]: I1124 17:45:02.512477 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xz9xq" event={"ID":"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510","Type":"ContainerDied","Data":"865fc6bbbfe63477f72e6511cdf09da548421a81f1f2d025c582821f266b8dda"} Nov 24 17:45:02 crc kubenswrapper[4808]: I1124 17:45:02.516584 4808 generic.go:334] "Generic (PLEG): container finished" podID="68962a01-d931-41b7-9492-a2d7fca41788" containerID="306ce16c3afa8e3fbcae31675bccc7902c4c9f4ff49f7ccfbfda0170eb0d50ee" exitCode=0 Nov 24 17:45:02 crc kubenswrapper[4808]: I1124 17:45:02.516609 4808 generic.go:334] "Generic (PLEG): container finished" podID="68962a01-d931-41b7-9492-a2d7fca41788" containerID="1a3d71f3643a1858ed2ba7b36ca80237b4ec90b8f96fab11420399c418d5ed46" exitCode=2 Nov 24 17:45:02 crc kubenswrapper[4808]: I1124 17:45:02.516618 4808 generic.go:334] "Generic (PLEG): container finished" podID="68962a01-d931-41b7-9492-a2d7fca41788" containerID="ad58ed120a79c69bdd7a2a466c479e5149fb3ea0c94b1ba71cef7b9334d65a1b" exitCode=0 Nov 24 17:45:02 crc kubenswrapper[4808]: I1124 17:45:02.516657 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68962a01-d931-41b7-9492-a2d7fca41788","Type":"ContainerDied","Data":"306ce16c3afa8e3fbcae31675bccc7902c4c9f4ff49f7ccfbfda0170eb0d50ee"} Nov 24 17:45:02 crc kubenswrapper[4808]: I1124 17:45:02.516680 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68962a01-d931-41b7-9492-a2d7fca41788","Type":"ContainerDied","Data":"1a3d71f3643a1858ed2ba7b36ca80237b4ec90b8f96fab11420399c418d5ed46"} Nov 24 17:45:02 crc kubenswrapper[4808]: I1124 17:45:02.516692 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68962a01-d931-41b7-9492-a2d7fca41788","Type":"ContainerDied","Data":"ad58ed120a79c69bdd7a2a466c479e5149fb3ea0c94b1ba71cef7b9334d65a1b"} Nov 24 17:45:02 crc kubenswrapper[4808]: I1124 17:45:02.518350 4808 generic.go:334] "Generic (PLEG): container finished" podID="71bc118d-05fa-4b8e-84ff-e8b9ad75de68" containerID="a3737b255ab02d41b6d3f0f2d908df7352943c3f48369d2be026048a147c587b" exitCode=0 Nov 24 17:45:02 crc kubenswrapper[4808]: I1124 17:45:02.518372 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" event={"ID":"71bc118d-05fa-4b8e-84ff-e8b9ad75de68","Type":"ContainerDied","Data":"a3737b255ab02d41b6d3f0f2d908df7352943c3f48369d2be026048a147c587b"} Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.182573 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.296170 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-combined-ca-bundle\") pod \"68962a01-d931-41b7-9492-a2d7fca41788\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.296311 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68962a01-d931-41b7-9492-a2d7fca41788-run-httpd\") pod \"68962a01-d931-41b7-9492-a2d7fca41788\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.296343 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2l47x\" (UniqueName: \"kubernetes.io/projected/68962a01-d931-41b7-9492-a2d7fca41788-kube-api-access-2l47x\") pod \"68962a01-d931-41b7-9492-a2d7fca41788\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.296429 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-scripts\") pod \"68962a01-d931-41b7-9492-a2d7fca41788\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.296462 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68962a01-d931-41b7-9492-a2d7fca41788-log-httpd\") pod \"68962a01-d931-41b7-9492-a2d7fca41788\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.296535 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-sg-core-conf-yaml\") pod \"68962a01-d931-41b7-9492-a2d7fca41788\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.296613 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-config-data\") pod \"68962a01-d931-41b7-9492-a2d7fca41788\" (UID: \"68962a01-d931-41b7-9492-a2d7fca41788\") " Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.296907 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68962a01-d931-41b7-9492-a2d7fca41788-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "68962a01-d931-41b7-9492-a2d7fca41788" (UID: "68962a01-d931-41b7-9492-a2d7fca41788"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.297174 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68962a01-d931-41b7-9492-a2d7fca41788-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "68962a01-d931-41b7-9492-a2d7fca41788" (UID: "68962a01-d931-41b7-9492-a2d7fca41788"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.297321 4808 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68962a01-d931-41b7-9492-a2d7fca41788-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.297351 4808 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68962a01-d931-41b7-9492-a2d7fca41788-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.310223 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-scripts" (OuterVolumeSpecName: "scripts") pod "68962a01-d931-41b7-9492-a2d7fca41788" (UID: "68962a01-d931-41b7-9492-a2d7fca41788"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.310258 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68962a01-d931-41b7-9492-a2d7fca41788-kube-api-access-2l47x" (OuterVolumeSpecName: "kube-api-access-2l47x") pod "68962a01-d931-41b7-9492-a2d7fca41788" (UID: "68962a01-d931-41b7-9492-a2d7fca41788"). InnerVolumeSpecName "kube-api-access-2l47x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.323961 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "68962a01-d931-41b7-9492-a2d7fca41788" (UID: "68962a01-d931-41b7-9492-a2d7fca41788"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.378198 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68962a01-d931-41b7-9492-a2d7fca41788" (UID: "68962a01-d931-41b7-9492-a2d7fca41788"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.391832 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-config-data" (OuterVolumeSpecName: "config-data") pod "68962a01-d931-41b7-9492-a2d7fca41788" (UID: "68962a01-d931-41b7-9492-a2d7fca41788"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.399365 4808 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.399400 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.399413 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.399424 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2l47x\" (UniqueName: \"kubernetes.io/projected/68962a01-d931-41b7-9492-a2d7fca41788-kube-api-access-2l47x\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.399440 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68962a01-d931-41b7-9492-a2d7fca41788-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.529760 4808 generic.go:334] "Generic (PLEG): container finished" podID="68962a01-d931-41b7-9492-a2d7fca41788" containerID="f0675119251252b15ac0b1e0a2f60a6e1291a0f5f9a91d7e0d02507675bc281e" exitCode=0 Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.529822 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.529843 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68962a01-d931-41b7-9492-a2d7fca41788","Type":"ContainerDied","Data":"f0675119251252b15ac0b1e0a2f60a6e1291a0f5f9a91d7e0d02507675bc281e"} Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.529902 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68962a01-d931-41b7-9492-a2d7fca41788","Type":"ContainerDied","Data":"218768293c045feaa7e0eaa19b643dff8fb80eca2f7fe872df0856854e283b40"} Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.529924 4808 scope.go:117] "RemoveContainer" containerID="306ce16c3afa8e3fbcae31675bccc7902c4c9f4ff49f7ccfbfda0170eb0d50ee" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.564150 4808 scope.go:117] "RemoveContainer" containerID="1a3d71f3643a1858ed2ba7b36ca80237b4ec90b8f96fab11420399c418d5ed46" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.585569 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.596910 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.597121 4808 scope.go:117] "RemoveContainer" containerID="ad58ed120a79c69bdd7a2a466c479e5149fb3ea0c94b1ba71cef7b9334d65a1b" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.615341 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:45:03 crc kubenswrapper[4808]: E1124 17:45:03.615842 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="ceilometer-central-agent" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.615864 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="ceilometer-central-agent" Nov 24 17:45:03 crc kubenswrapper[4808]: E1124 17:45:03.615888 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="proxy-httpd" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.615897 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="proxy-httpd" Nov 24 17:45:03 crc kubenswrapper[4808]: E1124 17:45:03.615916 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="sg-core" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.615925 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="sg-core" Nov 24 17:45:03 crc kubenswrapper[4808]: E1124 17:45:03.615958 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="ceilometer-notification-agent" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.615967 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="ceilometer-notification-agent" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.616239 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="proxy-httpd" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.616258 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="ceilometer-central-agent" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.616279 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="ceilometer-notification-agent" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.616293 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="68962a01-d931-41b7-9492-a2d7fca41788" containerName="sg-core" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.618533 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.621953 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.622179 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.628604 4808 scope.go:117] "RemoveContainer" containerID="f0675119251252b15ac0b1e0a2f60a6e1291a0f5f9a91d7e0d02507675bc281e" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.647079 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.656005 4808 scope.go:117] "RemoveContainer" containerID="306ce16c3afa8e3fbcae31675bccc7902c4c9f4ff49f7ccfbfda0170eb0d50ee" Nov 24 17:45:03 crc kubenswrapper[4808]: E1124 17:45:03.661425 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"306ce16c3afa8e3fbcae31675bccc7902c4c9f4ff49f7ccfbfda0170eb0d50ee\": container with ID starting with 306ce16c3afa8e3fbcae31675bccc7902c4c9f4ff49f7ccfbfda0170eb0d50ee not found: ID does not exist" containerID="306ce16c3afa8e3fbcae31675bccc7902c4c9f4ff49f7ccfbfda0170eb0d50ee" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.661467 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"306ce16c3afa8e3fbcae31675bccc7902c4c9f4ff49f7ccfbfda0170eb0d50ee"} err="failed to get container status \"306ce16c3afa8e3fbcae31675bccc7902c4c9f4ff49f7ccfbfda0170eb0d50ee\": rpc error: code = NotFound desc = could not find container \"306ce16c3afa8e3fbcae31675bccc7902c4c9f4ff49f7ccfbfda0170eb0d50ee\": container with ID starting with 306ce16c3afa8e3fbcae31675bccc7902c4c9f4ff49f7ccfbfda0170eb0d50ee not found: ID does not exist" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.661494 4808 scope.go:117] "RemoveContainer" containerID="1a3d71f3643a1858ed2ba7b36ca80237b4ec90b8f96fab11420399c418d5ed46" Nov 24 17:45:03 crc kubenswrapper[4808]: E1124 17:45:03.670679 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a3d71f3643a1858ed2ba7b36ca80237b4ec90b8f96fab11420399c418d5ed46\": container with ID starting with 1a3d71f3643a1858ed2ba7b36ca80237b4ec90b8f96fab11420399c418d5ed46 not found: ID does not exist" containerID="1a3d71f3643a1858ed2ba7b36ca80237b4ec90b8f96fab11420399c418d5ed46" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.670714 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a3d71f3643a1858ed2ba7b36ca80237b4ec90b8f96fab11420399c418d5ed46"} err="failed to get container status \"1a3d71f3643a1858ed2ba7b36ca80237b4ec90b8f96fab11420399c418d5ed46\": rpc error: code = NotFound desc = could not find container \"1a3d71f3643a1858ed2ba7b36ca80237b4ec90b8f96fab11420399c418d5ed46\": container with ID starting with 1a3d71f3643a1858ed2ba7b36ca80237b4ec90b8f96fab11420399c418d5ed46 not found: ID does not exist" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.670737 4808 scope.go:117] "RemoveContainer" containerID="ad58ed120a79c69bdd7a2a466c479e5149fb3ea0c94b1ba71cef7b9334d65a1b" Nov 24 17:45:03 crc kubenswrapper[4808]: E1124 17:45:03.671107 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad58ed120a79c69bdd7a2a466c479e5149fb3ea0c94b1ba71cef7b9334d65a1b\": container with ID starting with ad58ed120a79c69bdd7a2a466c479e5149fb3ea0c94b1ba71cef7b9334d65a1b not found: ID does not exist" containerID="ad58ed120a79c69bdd7a2a466c479e5149fb3ea0c94b1ba71cef7b9334d65a1b" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.671129 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad58ed120a79c69bdd7a2a466c479e5149fb3ea0c94b1ba71cef7b9334d65a1b"} err="failed to get container status \"ad58ed120a79c69bdd7a2a466c479e5149fb3ea0c94b1ba71cef7b9334d65a1b\": rpc error: code = NotFound desc = could not find container \"ad58ed120a79c69bdd7a2a466c479e5149fb3ea0c94b1ba71cef7b9334d65a1b\": container with ID starting with ad58ed120a79c69bdd7a2a466c479e5149fb3ea0c94b1ba71cef7b9334d65a1b not found: ID does not exist" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.671145 4808 scope.go:117] "RemoveContainer" containerID="f0675119251252b15ac0b1e0a2f60a6e1291a0f5f9a91d7e0d02507675bc281e" Nov 24 17:45:03 crc kubenswrapper[4808]: E1124 17:45:03.672194 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0675119251252b15ac0b1e0a2f60a6e1291a0f5f9a91d7e0d02507675bc281e\": container with ID starting with f0675119251252b15ac0b1e0a2f60a6e1291a0f5f9a91d7e0d02507675bc281e not found: ID does not exist" containerID="f0675119251252b15ac0b1e0a2f60a6e1291a0f5f9a91d7e0d02507675bc281e" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.672216 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0675119251252b15ac0b1e0a2f60a6e1291a0f5f9a91d7e0d02507675bc281e"} err="failed to get container status \"f0675119251252b15ac0b1e0a2f60a6e1291a0f5f9a91d7e0d02507675bc281e\": rpc error: code = NotFound desc = could not find container \"f0675119251252b15ac0b1e0a2f60a6e1291a0f5f9a91d7e0d02507675bc281e\": container with ID starting with f0675119251252b15ac0b1e0a2f60a6e1291a0f5f9a91d7e0d02507675bc281e not found: ID does not exist" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.703994 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djr4w\" (UniqueName: \"kubernetes.io/projected/77896689-86fa-4446-abd2-7c19ffde465c-kube-api-access-djr4w\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.704176 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77896689-86fa-4446-abd2-7c19ffde465c-run-httpd\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.704225 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.704300 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-scripts\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.704328 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77896689-86fa-4446-abd2-7c19ffde465c-log-httpd\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.704366 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-config-data\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.704481 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.807062 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.807543 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djr4w\" (UniqueName: \"kubernetes.io/projected/77896689-86fa-4446-abd2-7c19ffde465c-kube-api-access-djr4w\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.807615 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77896689-86fa-4446-abd2-7c19ffde465c-run-httpd\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.807639 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.807698 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-scripts\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.807723 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77896689-86fa-4446-abd2-7c19ffde465c-log-httpd\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.807745 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-config-data\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.810657 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77896689-86fa-4446-abd2-7c19ffde465c-run-httpd\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.811862 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77896689-86fa-4446-abd2-7c19ffde465c-log-httpd\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.812600 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.817696 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.818153 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-scripts\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.819664 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-config-data\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.832829 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djr4w\" (UniqueName: \"kubernetes.io/projected/77896689-86fa-4446-abd2-7c19ffde465c-kube-api-access-djr4w\") pod \"ceilometer-0\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " pod="openstack/ceilometer-0" Nov 24 17:45:03 crc kubenswrapper[4808]: I1124 17:45:03.944856 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.063942 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.068919 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.114090 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9gtk\" (UniqueName: \"kubernetes.io/projected/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-kube-api-access-z9gtk\") pod \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.114344 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-config-data\") pod \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.114579 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-secret-volume\") pod \"71bc118d-05fa-4b8e-84ff-e8b9ad75de68\" (UID: \"71bc118d-05fa-4b8e-84ff-e8b9ad75de68\") " Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.114690 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-scripts\") pod \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.114833 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-combined-ca-bundle\") pod \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\" (UID: \"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510\") " Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.114939 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-config-volume\") pod \"71bc118d-05fa-4b8e-84ff-e8b9ad75de68\" (UID: \"71bc118d-05fa-4b8e-84ff-e8b9ad75de68\") " Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.115077 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9qwz\" (UniqueName: \"kubernetes.io/projected/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-kube-api-access-l9qwz\") pod \"71bc118d-05fa-4b8e-84ff-e8b9ad75de68\" (UID: \"71bc118d-05fa-4b8e-84ff-e8b9ad75de68\") " Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.116499 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-config-volume" (OuterVolumeSpecName: "config-volume") pod "71bc118d-05fa-4b8e-84ff-e8b9ad75de68" (UID: "71bc118d-05fa-4b8e-84ff-e8b9ad75de68"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.125559 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "71bc118d-05fa-4b8e-84ff-e8b9ad75de68" (UID: "71bc118d-05fa-4b8e-84ff-e8b9ad75de68"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.127418 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-kube-api-access-z9gtk" (OuterVolumeSpecName: "kube-api-access-z9gtk") pod "bd6dfdb4-e6c9-4302-bdf5-cbb81348a510" (UID: "bd6dfdb4-e6c9-4302-bdf5-cbb81348a510"). InnerVolumeSpecName "kube-api-access-z9gtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.131133 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-kube-api-access-l9qwz" (OuterVolumeSpecName: "kube-api-access-l9qwz") pod "71bc118d-05fa-4b8e-84ff-e8b9ad75de68" (UID: "71bc118d-05fa-4b8e-84ff-e8b9ad75de68"). InnerVolumeSpecName "kube-api-access-l9qwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.141650 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-scripts" (OuterVolumeSpecName: "scripts") pod "bd6dfdb4-e6c9-4302-bdf5-cbb81348a510" (UID: "bd6dfdb4-e6c9-4302-bdf5-cbb81348a510"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.159991 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd6dfdb4-e6c9-4302-bdf5-cbb81348a510" (UID: "bd6dfdb4-e6c9-4302-bdf5-cbb81348a510"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.178077 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-config-data" (OuterVolumeSpecName: "config-data") pod "bd6dfdb4-e6c9-4302-bdf5-cbb81348a510" (UID: "bd6dfdb4-e6c9-4302-bdf5-cbb81348a510"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.216970 4808 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.217005 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.217039 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.217051 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.217064 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9qwz\" (UniqueName: \"kubernetes.io/projected/71bc118d-05fa-4b8e-84ff-e8b9ad75de68-kube-api-access-l9qwz\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.217076 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9gtk\" (UniqueName: \"kubernetes.io/projected/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-kube-api-access-z9gtk\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.217086 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.357981 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68962a01-d931-41b7-9492-a2d7fca41788" path="/var/lib/kubelet/pods/68962a01-d931-41b7-9492-a2d7fca41788/volumes" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.407651 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:45:04 crc kubenswrapper[4808]: W1124 17:45:04.412138 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77896689_86fa_4446_abd2_7c19ffde465c.slice/crio-89ebf06e7247655950bc18445ce659362993bca483976296ca7340537fdf5877 WatchSource:0}: Error finding container 89ebf06e7247655950bc18445ce659362993bca483976296ca7340537fdf5877: Status 404 returned error can't find the container with id 89ebf06e7247655950bc18445ce659362993bca483976296ca7340537fdf5877 Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.551251 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.552060 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp" event={"ID":"71bc118d-05fa-4b8e-84ff-e8b9ad75de68","Type":"ContainerDied","Data":"d3b99ec44910f20f22dc01f6ff3844bd114c54b01e818e41a39f7a3af09123f7"} Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.552087 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3b99ec44910f20f22dc01f6ff3844bd114c54b01e818e41a39f7a3af09123f7" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.553308 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77896689-86fa-4446-abd2-7c19ffde465c","Type":"ContainerStarted","Data":"89ebf06e7247655950bc18445ce659362993bca483976296ca7340537fdf5877"} Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.558587 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xz9xq" event={"ID":"bd6dfdb4-e6c9-4302-bdf5-cbb81348a510","Type":"ContainerDied","Data":"d6c6b7fb8076934f76a611688933217bc8881288ad55eabdcb622927d196fb26"} Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.558630 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6c6b7fb8076934f76a611688933217bc8881288ad55eabdcb622927d196fb26" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.558692 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xz9xq" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.641758 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 17:45:04 crc kubenswrapper[4808]: E1124 17:45:04.642263 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd6dfdb4-e6c9-4302-bdf5-cbb81348a510" containerName="nova-cell0-conductor-db-sync" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.642292 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd6dfdb4-e6c9-4302-bdf5-cbb81348a510" containerName="nova-cell0-conductor-db-sync" Nov 24 17:45:04 crc kubenswrapper[4808]: E1124 17:45:04.642310 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71bc118d-05fa-4b8e-84ff-e8b9ad75de68" containerName="collect-profiles" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.642316 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="71bc118d-05fa-4b8e-84ff-e8b9ad75de68" containerName="collect-profiles" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.642501 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="71bc118d-05fa-4b8e-84ff-e8b9ad75de68" containerName="collect-profiles" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.642543 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd6dfdb4-e6c9-4302-bdf5-cbb81348a510" containerName="nova-cell0-conductor-db-sync" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.643277 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.646564 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-b5s49" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.646784 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.658997 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.724047 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c4e90b9-4ce7-4931-922f-ce473cc9d5ed-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1c4e90b9-4ce7-4931-922f-ce473cc9d5ed\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.724452 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9tt8\" (UniqueName: \"kubernetes.io/projected/1c4e90b9-4ce7-4931-922f-ce473cc9d5ed-kube-api-access-s9tt8\") pod \"nova-cell0-conductor-0\" (UID: \"1c4e90b9-4ce7-4931-922f-ce473cc9d5ed\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.724496 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c4e90b9-4ce7-4931-922f-ce473cc9d5ed-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1c4e90b9-4ce7-4931-922f-ce473cc9d5ed\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.825897 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9tt8\" (UniqueName: \"kubernetes.io/projected/1c4e90b9-4ce7-4931-922f-ce473cc9d5ed-kube-api-access-s9tt8\") pod \"nova-cell0-conductor-0\" (UID: \"1c4e90b9-4ce7-4931-922f-ce473cc9d5ed\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.825947 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c4e90b9-4ce7-4931-922f-ce473cc9d5ed-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1c4e90b9-4ce7-4931-922f-ce473cc9d5ed\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.826083 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c4e90b9-4ce7-4931-922f-ce473cc9d5ed-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1c4e90b9-4ce7-4931-922f-ce473cc9d5ed\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.832235 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c4e90b9-4ce7-4931-922f-ce473cc9d5ed-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1c4e90b9-4ce7-4931-922f-ce473cc9d5ed\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.832326 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c4e90b9-4ce7-4931-922f-ce473cc9d5ed-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1c4e90b9-4ce7-4931-922f-ce473cc9d5ed\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.849493 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9tt8\" (UniqueName: \"kubernetes.io/projected/1c4e90b9-4ce7-4931-922f-ce473cc9d5ed-kube-api-access-s9tt8\") pod \"nova-cell0-conductor-0\" (UID: \"1c4e90b9-4ce7-4931-922f-ce473cc9d5ed\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:45:04 crc kubenswrapper[4808]: I1124 17:45:04.964052 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 17:45:05 crc kubenswrapper[4808]: I1124 17:45:05.249711 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 17:45:05 crc kubenswrapper[4808]: W1124 17:45:05.257372 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c4e90b9_4ce7_4931_922f_ce473cc9d5ed.slice/crio-ca8337fb9a2b1c5c17a8257f28f981454ffe3ed9d07d716c40f04560cfae21f5 WatchSource:0}: Error finding container ca8337fb9a2b1c5c17a8257f28f981454ffe3ed9d07d716c40f04560cfae21f5: Status 404 returned error can't find the container with id ca8337fb9a2b1c5c17a8257f28f981454ffe3ed9d07d716c40f04560cfae21f5 Nov 24 17:45:05 crc kubenswrapper[4808]: I1124 17:45:05.572587 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1c4e90b9-4ce7-4931-922f-ce473cc9d5ed","Type":"ContainerStarted","Data":"f9e5a2a35156b0b7491636e79c033d4e0ef77389f9375c22d208df1baabeaed0"} Nov 24 17:45:05 crc kubenswrapper[4808]: I1124 17:45:05.573055 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1c4e90b9-4ce7-4931-922f-ce473cc9d5ed","Type":"ContainerStarted","Data":"ca8337fb9a2b1c5c17a8257f28f981454ffe3ed9d07d716c40f04560cfae21f5"} Nov 24 17:45:05 crc kubenswrapper[4808]: I1124 17:45:05.573175 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 17:45:05 crc kubenswrapper[4808]: I1124 17:45:05.575041 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77896689-86fa-4446-abd2-7c19ffde465c","Type":"ContainerStarted","Data":"6b77fe1160b9842403f786da70a60e66398685601ab26dfdefd3d1c998398a64"} Nov 24 17:45:05 crc kubenswrapper[4808]: I1124 17:45:05.595967 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.5959479760000002 podStartE2EDuration="1.595947976s" podCreationTimestamp="2025-11-24 17:45:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:45:05.59404146 +0000 UTC m=+1098.191709282" watchObservedRunningTime="2025-11-24 17:45:05.595947976 +0000 UTC m=+1098.193615778" Nov 24 17:45:06 crc kubenswrapper[4808]: I1124 17:45:06.584990 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77896689-86fa-4446-abd2-7c19ffde465c","Type":"ContainerStarted","Data":"9e7e050c43538878ac97b5c541902c1729903db81dab306533c5bb4939ccb308"} Nov 24 17:45:06 crc kubenswrapper[4808]: I1124 17:45:06.586295 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77896689-86fa-4446-abd2-7c19ffde465c","Type":"ContainerStarted","Data":"1870350250d37b6b52d3bae0f6aa27541a7b6010f2c0de39dff35756d4a59b2c"} Nov 24 17:45:09 crc kubenswrapper[4808]: I1124 17:45:09.615313 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77896689-86fa-4446-abd2-7c19ffde465c","Type":"ContainerStarted","Data":"f4ebc9ca0a99ca9f376a6f519e935c3ad22eb0a898b1d31ec7073c30f90aa108"} Nov 24 17:45:09 crc kubenswrapper[4808]: I1124 17:45:09.616118 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 17:45:09 crc kubenswrapper[4808]: I1124 17:45:09.646118 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.579494968 podStartE2EDuration="6.64609822s" podCreationTimestamp="2025-11-24 17:45:03 +0000 UTC" firstStartedPulling="2025-11-24 17:45:04.4142403 +0000 UTC m=+1097.011908102" lastFinishedPulling="2025-11-24 17:45:08.480843552 +0000 UTC m=+1101.078511354" observedRunningTime="2025-11-24 17:45:09.636409013 +0000 UTC m=+1102.234076825" watchObservedRunningTime="2025-11-24 17:45:09.64609822 +0000 UTC m=+1102.243766032" Nov 24 17:45:14 crc kubenswrapper[4808]: I1124 17:45:14.991114 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.403292 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-lfpxw"] Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.404753 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.406592 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.409087 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.414582 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lfpxw"] Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.517667 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lfpxw\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.517714 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-scripts\") pod \"nova-cell0-cell-mapping-lfpxw\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.517762 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-config-data\") pod \"nova-cell0-cell-mapping-lfpxw\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.517904 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28s6k\" (UniqueName: \"kubernetes.io/projected/1ff6c89f-2920-4894-8036-bbf316ecd48b-kube-api-access-28s6k\") pod \"nova-cell0-cell-mapping-lfpxw\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.542416 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.545750 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.550691 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.553576 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.617154 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.618662 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.619501 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28s6k\" (UniqueName: \"kubernetes.io/projected/1ff6c89f-2920-4894-8036-bbf316ecd48b-kube-api-access-28s6k\") pod \"nova-cell0-cell-mapping-lfpxw\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.620519 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lfpxw\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.620566 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-scripts\") pod \"nova-cell0-cell-mapping-lfpxw\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.620615 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-config-data\") pod \"nova-cell0-cell-mapping-lfpxw\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.622140 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.626711 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-config-data\") pod \"nova-cell0-cell-mapping-lfpxw\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.629394 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.641630 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lfpxw\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.655572 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-scripts\") pod \"nova-cell0-cell-mapping-lfpxw\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.661585 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28s6k\" (UniqueName: \"kubernetes.io/projected/1ff6c89f-2920-4894-8036-bbf316ecd48b-kube-api-access-28s6k\") pod \"nova-cell0-cell-mapping-lfpxw\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.670704 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.672413 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.676054 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.720639 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.724661 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82969a14-94f6-4c08-8d47-b65543a52e80-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"82969a14-94f6-4c08-8d47-b65543a52e80\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.724706 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-logs\") pod \"nova-api-0\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " pod="openstack/nova-api-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.724737 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46fxj\" (UniqueName: \"kubernetes.io/projected/82969a14-94f6-4c08-8d47-b65543a52e80-kube-api-access-46fxj\") pod \"nova-cell1-novncproxy-0\" (UID: \"82969a14-94f6-4c08-8d47-b65543a52e80\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.724773 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8647155b-5589-4daa-b1a8-5bbffcc6e3f4\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.724800 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-config-data\") pod \"nova-api-0\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " pod="openstack/nova-api-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.724822 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-config-data\") pod \"nova-scheduler-0\" (UID: \"8647155b-5589-4daa-b1a8-5bbffcc6e3f4\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.724853 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjb8z\" (UniqueName: \"kubernetes.io/projected/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-kube-api-access-pjb8z\") pod \"nova-api-0\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " pod="openstack/nova-api-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.724883 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82969a14-94f6-4c08-8d47-b65543a52e80-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"82969a14-94f6-4c08-8d47-b65543a52e80\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.724932 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnkpz\" (UniqueName: \"kubernetes.io/projected/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-kube-api-access-bnkpz\") pod \"nova-scheduler-0\" (UID: \"8647155b-5589-4daa-b1a8-5bbffcc6e3f4\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.724952 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " pod="openstack/nova-api-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.725155 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.819831 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.822304 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.828085 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-config-data\") pod \"nova-scheduler-0\" (UID: \"8647155b-5589-4daa-b1a8-5bbffcc6e3f4\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.828173 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b62eafcb-c14a-442e-b8ce-231bcbbb434e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.828224 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjb8z\" (UniqueName: \"kubernetes.io/projected/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-kube-api-access-pjb8z\") pod \"nova-api-0\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " pod="openstack/nova-api-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.828283 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82969a14-94f6-4c08-8d47-b65543a52e80-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"82969a14-94f6-4c08-8d47-b65543a52e80\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.828348 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnkpz\" (UniqueName: \"kubernetes.io/projected/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-kube-api-access-bnkpz\") pod \"nova-scheduler-0\" (UID: \"8647155b-5589-4daa-b1a8-5bbffcc6e3f4\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.828387 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " pod="openstack/nova-api-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.828414 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82969a14-94f6-4c08-8d47-b65543a52e80-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"82969a14-94f6-4c08-8d47-b65543a52e80\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.828448 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b62eafcb-c14a-442e-b8ce-231bcbbb434e-logs\") pod \"nova-metadata-0\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.828482 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-logs\") pod \"nova-api-0\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " pod="openstack/nova-api-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.828522 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv5t4\" (UniqueName: \"kubernetes.io/projected/b62eafcb-c14a-442e-b8ce-231bcbbb434e-kube-api-access-bv5t4\") pod \"nova-metadata-0\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.828552 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46fxj\" (UniqueName: \"kubernetes.io/projected/82969a14-94f6-4c08-8d47-b65543a52e80-kube-api-access-46fxj\") pod \"nova-cell1-novncproxy-0\" (UID: \"82969a14-94f6-4c08-8d47-b65543a52e80\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.828597 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b62eafcb-c14a-442e-b8ce-231bcbbb434e-config-data\") pod \"nova-metadata-0\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.828639 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8647155b-5589-4daa-b1a8-5bbffcc6e3f4\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.828692 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-config-data\") pod \"nova-api-0\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " pod="openstack/nova-api-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.830251 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.830823 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-logs\") pod \"nova-api-0\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " pod="openstack/nova-api-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.835625 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-config-data\") pod \"nova-api-0\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " pod="openstack/nova-api-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.837265 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82969a14-94f6-4c08-8d47-b65543a52e80-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"82969a14-94f6-4c08-8d47-b65543a52e80\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.837764 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " pod="openstack/nova-api-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.839827 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-config-data\") pod \"nova-scheduler-0\" (UID: \"8647155b-5589-4daa-b1a8-5bbffcc6e3f4\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.848712 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82969a14-94f6-4c08-8d47-b65543a52e80-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"82969a14-94f6-4c08-8d47-b65543a52e80\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.850001 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.851415 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8647155b-5589-4daa-b1a8-5bbffcc6e3f4\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.853834 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjb8z\" (UniqueName: \"kubernetes.io/projected/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-kube-api-access-pjb8z\") pod \"nova-api-0\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " pod="openstack/nova-api-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.860706 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnkpz\" (UniqueName: \"kubernetes.io/projected/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-kube-api-access-bnkpz\") pod \"nova-scheduler-0\" (UID: \"8647155b-5589-4daa-b1a8-5bbffcc6e3f4\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.868716 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46fxj\" (UniqueName: \"kubernetes.io/projected/82969a14-94f6-4c08-8d47-b65543a52e80-kube-api-access-46fxj\") pod \"nova-cell1-novncproxy-0\" (UID: \"82969a14-94f6-4c08-8d47-b65543a52e80\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.870687 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.910073 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.929993 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b62eafcb-c14a-442e-b8ce-231bcbbb434e-logs\") pod \"nova-metadata-0\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.930074 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv5t4\" (UniqueName: \"kubernetes.io/projected/b62eafcb-c14a-442e-b8ce-231bcbbb434e-kube-api-access-bv5t4\") pod \"nova-metadata-0\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.930116 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b62eafcb-c14a-442e-b8ce-231bcbbb434e-config-data\") pod \"nova-metadata-0\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.930268 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b62eafcb-c14a-442e-b8ce-231bcbbb434e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.931218 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b62eafcb-c14a-442e-b8ce-231bcbbb434e-logs\") pod \"nova-metadata-0\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.936936 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b62eafcb-c14a-442e-b8ce-231bcbbb434e-config-data\") pod \"nova-metadata-0\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.937775 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b62eafcb-c14a-442e-b8ce-231bcbbb434e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.946394 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-s2859"] Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.947916 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.968924 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-s2859"] Nov 24 17:45:15 crc kubenswrapper[4808]: I1124 17:45:15.970106 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv5t4\" (UniqueName: \"kubernetes.io/projected/b62eafcb-c14a-442e-b8ce-231bcbbb434e-kube-api-access-bv5t4\") pod \"nova-metadata-0\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.050082 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.133888 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.133944 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-config\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.133990 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.134054 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc5lx\" (UniqueName: \"kubernetes.io/projected/6b049d97-3af4-4771-8be7-ee521533bce9-kube-api-access-kc5lx\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.134097 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.134137 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-dns-svc\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.227950 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.250008 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.250376 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-config\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.250489 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.250636 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc5lx\" (UniqueName: \"kubernetes.io/projected/6b049d97-3af4-4771-8be7-ee521533bce9-kube-api-access-kc5lx\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.250840 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.250889 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-dns-svc\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.251351 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.258837 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-config\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.261189 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.261417 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-dns-svc\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.261586 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.269386 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc5lx\" (UniqueName: \"kubernetes.io/projected/6b049d97-3af4-4771-8be7-ee521533bce9-kube-api-access-kc5lx\") pod \"dnsmasq-dns-757b4f8459-s2859\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.290983 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lfpxw"] Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.295918 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:16 crc kubenswrapper[4808]: W1124 17:45:16.311144 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ff6c89f_2920_4894_8036_bbf316ecd48b.slice/crio-3e4783e062ee65c957d20483bcaba76a6f8872edfd2407c8709d0619bd1dc770 WatchSource:0}: Error finding container 3e4783e062ee65c957d20483bcaba76a6f8872edfd2407c8709d0619bd1dc770: Status 404 returned error can't find the container with id 3e4783e062ee65c957d20483bcaba76a6f8872edfd2407c8709d0619bd1dc770 Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.497343 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:45:16 crc kubenswrapper[4808]: W1124 17:45:16.507526 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8647155b_5589_4daa_b1a8_5bbffcc6e3f4.slice/crio-e9190293f873ef38dcf871c397b6dfe8aec3c4c3515abff32db6ec7a5a3daefe WatchSource:0}: Error finding container e9190293f873ef38dcf871c397b6dfe8aec3c4c3515abff32db6ec7a5a3daefe: Status 404 returned error can't find the container with id e9190293f873ef38dcf871c397b6dfe8aec3c4c3515abff32db6ec7a5a3daefe Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.554779 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.566095 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.710117 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8647155b-5589-4daa-b1a8-5bbffcc6e3f4","Type":"ContainerStarted","Data":"e9190293f873ef38dcf871c397b6dfe8aec3c4c3515abff32db6ec7a5a3daefe"} Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.712292 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rfs5s"] Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.713546 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.716412 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.716467 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.717173 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599","Type":"ContainerStarted","Data":"484b416078ac71035dc34e3bae43ef0532fa8c7375d5dde9004f848103812867"} Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.720942 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lfpxw" event={"ID":"1ff6c89f-2920-4894-8036-bbf316ecd48b","Type":"ContainerStarted","Data":"522929826dc3bfc36bdc89b2ee5bc6c4d9d79ae11afa1c70662522d45d6b603d"} Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.720990 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lfpxw" event={"ID":"1ff6c89f-2920-4894-8036-bbf316ecd48b","Type":"ContainerStarted","Data":"3e4783e062ee65c957d20483bcaba76a6f8872edfd2407c8709d0619bd1dc770"} Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.724583 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"82969a14-94f6-4c08-8d47-b65543a52e80","Type":"ContainerStarted","Data":"9b80a6144bf96bea554448f695440df0a3edebb118a73ecdb9ff917470d469cf"} Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.731848 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rfs5s"] Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.751033 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-lfpxw" podStartSLOduration=1.750994352 podStartE2EDuration="1.750994352s" podCreationTimestamp="2025-11-24 17:45:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:45:16.745811894 +0000 UTC m=+1109.343479696" watchObservedRunningTime="2025-11-24 17:45:16.750994352 +0000 UTC m=+1109.348662154" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.853415 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.870738 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qnz9\" (UniqueName: \"kubernetes.io/projected/1373d3ee-b91b-413e-af94-40d6c55c1917-kube-api-access-8qnz9\") pod \"nova-cell1-conductor-db-sync-rfs5s\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.870834 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-scripts\") pod \"nova-cell1-conductor-db-sync-rfs5s\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.870921 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rfs5s\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.870957 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-config-data\") pod \"nova-cell1-conductor-db-sync-rfs5s\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.914556 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-s2859"] Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.972978 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-scripts\") pod \"nova-cell1-conductor-db-sync-rfs5s\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.973124 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rfs5s\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.973160 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-config-data\") pod \"nova-cell1-conductor-db-sync-rfs5s\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.973299 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qnz9\" (UniqueName: \"kubernetes.io/projected/1373d3ee-b91b-413e-af94-40d6c55c1917-kube-api-access-8qnz9\") pod \"nova-cell1-conductor-db-sync-rfs5s\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.978342 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rfs5s\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.981259 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-scripts\") pod \"nova-cell1-conductor-db-sync-rfs5s\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.981973 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-config-data\") pod \"nova-cell1-conductor-db-sync-rfs5s\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:16 crc kubenswrapper[4808]: I1124 17:45:16.991030 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qnz9\" (UniqueName: \"kubernetes.io/projected/1373d3ee-b91b-413e-af94-40d6c55c1917-kube-api-access-8qnz9\") pod \"nova-cell1-conductor-db-sync-rfs5s\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:17 crc kubenswrapper[4808]: I1124 17:45:17.044560 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:17 crc kubenswrapper[4808]: I1124 17:45:17.536930 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rfs5s"] Nov 24 17:45:17 crc kubenswrapper[4808]: W1124 17:45:17.555905 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1373d3ee_b91b_413e_af94_40d6c55c1917.slice/crio-661df439ac3ccacbc182cf9ad6947dee5f55f1e84595515e43259fd592a64257 WatchSource:0}: Error finding container 661df439ac3ccacbc182cf9ad6947dee5f55f1e84595515e43259fd592a64257: Status 404 returned error can't find the container with id 661df439ac3ccacbc182cf9ad6947dee5f55f1e84595515e43259fd592a64257 Nov 24 17:45:17 crc kubenswrapper[4808]: I1124 17:45:17.752455 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b62eafcb-c14a-442e-b8ce-231bcbbb434e","Type":"ContainerStarted","Data":"88b843e5d937b83f6e16fd08a6d47a1e888d7e9a13eb4c82c9744ab21cf54ff2"} Nov 24 17:45:17 crc kubenswrapper[4808]: I1124 17:45:17.758072 4808 generic.go:334] "Generic (PLEG): container finished" podID="6b049d97-3af4-4771-8be7-ee521533bce9" containerID="8d33c1dbfc3c22434743ccbb9ae13a0f8b553a3967a885fe22b9b6f72460f6ff" exitCode=0 Nov 24 17:45:17 crc kubenswrapper[4808]: I1124 17:45:17.758157 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-s2859" event={"ID":"6b049d97-3af4-4771-8be7-ee521533bce9","Type":"ContainerDied","Data":"8d33c1dbfc3c22434743ccbb9ae13a0f8b553a3967a885fe22b9b6f72460f6ff"} Nov 24 17:45:17 crc kubenswrapper[4808]: I1124 17:45:17.758189 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-s2859" event={"ID":"6b049d97-3af4-4771-8be7-ee521533bce9","Type":"ContainerStarted","Data":"a561329c8ea54407910f27a50aefc674d5f8d5566e192fecc0357879cc024a46"} Nov 24 17:45:17 crc kubenswrapper[4808]: I1124 17:45:17.762211 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rfs5s" event={"ID":"1373d3ee-b91b-413e-af94-40d6c55c1917","Type":"ContainerStarted","Data":"661df439ac3ccacbc182cf9ad6947dee5f55f1e84595515e43259fd592a64257"} Nov 24 17:45:19 crc kubenswrapper[4808]: I1124 17:45:19.503184 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:45:19 crc kubenswrapper[4808]: I1124 17:45:19.581693 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:45:20 crc kubenswrapper[4808]: I1124 17:45:20.819826 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rfs5s" event={"ID":"1373d3ee-b91b-413e-af94-40d6c55c1917","Type":"ContainerStarted","Data":"82ba366ba8d51289ea732089d13e7cadb0dd0c2c292e82a6b1e301f97a7ae5ab"} Nov 24 17:45:20 crc kubenswrapper[4808]: I1124 17:45:20.840075 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-rfs5s" podStartSLOduration=4.840054433 podStartE2EDuration="4.840054433s" podCreationTimestamp="2025-11-24 17:45:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:45:20.838372455 +0000 UTC m=+1113.436040257" watchObservedRunningTime="2025-11-24 17:45:20.840054433 +0000 UTC m=+1113.437722235" Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.832121 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"82969a14-94f6-4c08-8d47-b65543a52e80","Type":"ContainerStarted","Data":"2b6e9e6927c329fd49c6eb9a7915a78599d7b187b0e36d07351ade5f98c184fe"} Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.832265 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="82969a14-94f6-4c08-8d47-b65543a52e80" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2b6e9e6927c329fd49c6eb9a7915a78599d7b187b0e36d07351ade5f98c184fe" gracePeriod=30 Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.834552 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8647155b-5589-4daa-b1a8-5bbffcc6e3f4","Type":"ContainerStarted","Data":"f0fa9927bcfe60084ac3154126f4def663f8221c09275f1c8bdcb7c2519764d6"} Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.844097 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599","Type":"ContainerStarted","Data":"84f7d4e24d959ce8f11afb76b8220e2fb780828e8701136618a2be208a1d5dfa"} Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.844159 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599","Type":"ContainerStarted","Data":"07ce607292a080bb811b46df930b8b0039a46023a74a45bff26944dc5fa5c773"} Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.846509 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b62eafcb-c14a-442e-b8ce-231bcbbb434e","Type":"ContainerStarted","Data":"c74d6e32d7fe951e38a94f91562a264d2be51496678b451af2e176ad4583ffe0"} Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.846544 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b62eafcb-c14a-442e-b8ce-231bcbbb434e","Type":"ContainerStarted","Data":"b114fd29e32611d2048b0bd3384298dc3a6e8bfd6fc5905cb04b5627cfa39155"} Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.846652 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b62eafcb-c14a-442e-b8ce-231bcbbb434e" containerName="nova-metadata-log" containerID="cri-o://b114fd29e32611d2048b0bd3384298dc3a6e8bfd6fc5905cb04b5627cfa39155" gracePeriod=30 Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.846987 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b62eafcb-c14a-442e-b8ce-231bcbbb434e" containerName="nova-metadata-metadata" containerID="cri-o://c74d6e32d7fe951e38a94f91562a264d2be51496678b451af2e176ad4583ffe0" gracePeriod=30 Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.853242 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-s2859" event={"ID":"6b049d97-3af4-4771-8be7-ee521533bce9","Type":"ContainerStarted","Data":"d6f0b7494cf8448bee489c7f4d06af24466bf0a30c36a8e5227fa2ea327f62e1"} Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.853354 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.867435 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.934166243 podStartE2EDuration="6.867412724s" podCreationTimestamp="2025-11-24 17:45:15 +0000 UTC" firstStartedPulling="2025-11-24 17:45:16.579527055 +0000 UTC m=+1109.177194857" lastFinishedPulling="2025-11-24 17:45:20.512773536 +0000 UTC m=+1113.110441338" observedRunningTime="2025-11-24 17:45:21.854397112 +0000 UTC m=+1114.452064934" watchObservedRunningTime="2025-11-24 17:45:21.867412724 +0000 UTC m=+1114.465080526" Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.888126 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-s2859" podStartSLOduration=6.888104345 podStartE2EDuration="6.888104345s" podCreationTimestamp="2025-11-24 17:45:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:45:21.885034267 +0000 UTC m=+1114.482702079" watchObservedRunningTime="2025-11-24 17:45:21.888104345 +0000 UTC m=+1114.485772147" Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.910870 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.214727726 podStartE2EDuration="6.910853965s" podCreationTimestamp="2025-11-24 17:45:15 +0000 UTC" firstStartedPulling="2025-11-24 17:45:16.850821453 +0000 UTC m=+1109.448489255" lastFinishedPulling="2025-11-24 17:45:20.546947692 +0000 UTC m=+1113.144615494" observedRunningTime="2025-11-24 17:45:21.902312101 +0000 UTC m=+1114.499979903" watchObservedRunningTime="2025-11-24 17:45:21.910853965 +0000 UTC m=+1114.508521767" Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.990671 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.994719973 podStartE2EDuration="6.990646544s" podCreationTimestamp="2025-11-24 17:45:15 +0000 UTC" firstStartedPulling="2025-11-24 17:45:16.515521057 +0000 UTC m=+1109.113188859" lastFinishedPulling="2025-11-24 17:45:20.511447628 +0000 UTC m=+1113.109115430" observedRunningTime="2025-11-24 17:45:21.924518495 +0000 UTC m=+1114.522186297" watchObservedRunningTime="2025-11-24 17:45:21.990646544 +0000 UTC m=+1114.588314346" Nov 24 17:45:21 crc kubenswrapper[4808]: I1124 17:45:21.998062 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.064047682 podStartE2EDuration="6.998041475s" podCreationTimestamp="2025-11-24 17:45:15 +0000 UTC" firstStartedPulling="2025-11-24 17:45:16.578424973 +0000 UTC m=+1109.176092775" lastFinishedPulling="2025-11-24 17:45:20.512418766 +0000 UTC m=+1113.110086568" observedRunningTime="2025-11-24 17:45:21.958616289 +0000 UTC m=+1114.556284111" watchObservedRunningTime="2025-11-24 17:45:21.998041475 +0000 UTC m=+1114.595709277" Nov 24 17:45:22 crc kubenswrapper[4808]: I1124 17:45:22.867842 4808 generic.go:334] "Generic (PLEG): container finished" podID="b62eafcb-c14a-442e-b8ce-231bcbbb434e" containerID="c74d6e32d7fe951e38a94f91562a264d2be51496678b451af2e176ad4583ffe0" exitCode=0 Nov 24 17:45:22 crc kubenswrapper[4808]: I1124 17:45:22.867880 4808 generic.go:334] "Generic (PLEG): container finished" podID="b62eafcb-c14a-442e-b8ce-231bcbbb434e" containerID="b114fd29e32611d2048b0bd3384298dc3a6e8bfd6fc5905cb04b5627cfa39155" exitCode=143 Nov 24 17:45:22 crc kubenswrapper[4808]: I1124 17:45:22.867923 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b62eafcb-c14a-442e-b8ce-231bcbbb434e","Type":"ContainerDied","Data":"c74d6e32d7fe951e38a94f91562a264d2be51496678b451af2e176ad4583ffe0"} Nov 24 17:45:22 crc kubenswrapper[4808]: I1124 17:45:22.867982 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b62eafcb-c14a-442e-b8ce-231bcbbb434e","Type":"ContainerDied","Data":"b114fd29e32611d2048b0bd3384298dc3a6e8bfd6fc5905cb04b5627cfa39155"} Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.021333 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.201299 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b62eafcb-c14a-442e-b8ce-231bcbbb434e-logs\") pod \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.201355 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b62eafcb-c14a-442e-b8ce-231bcbbb434e-config-data\") pod \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.201484 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b62eafcb-c14a-442e-b8ce-231bcbbb434e-combined-ca-bundle\") pod \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.201608 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv5t4\" (UniqueName: \"kubernetes.io/projected/b62eafcb-c14a-442e-b8ce-231bcbbb434e-kube-api-access-bv5t4\") pod \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\" (UID: \"b62eafcb-c14a-442e-b8ce-231bcbbb434e\") " Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.202291 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b62eafcb-c14a-442e-b8ce-231bcbbb434e-logs" (OuterVolumeSpecName: "logs") pod "b62eafcb-c14a-442e-b8ce-231bcbbb434e" (UID: "b62eafcb-c14a-442e-b8ce-231bcbbb434e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.209272 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b62eafcb-c14a-442e-b8ce-231bcbbb434e-kube-api-access-bv5t4" (OuterVolumeSpecName: "kube-api-access-bv5t4") pod "b62eafcb-c14a-442e-b8ce-231bcbbb434e" (UID: "b62eafcb-c14a-442e-b8ce-231bcbbb434e"). InnerVolumeSpecName "kube-api-access-bv5t4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.230235 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b62eafcb-c14a-442e-b8ce-231bcbbb434e-config-data" (OuterVolumeSpecName: "config-data") pod "b62eafcb-c14a-442e-b8ce-231bcbbb434e" (UID: "b62eafcb-c14a-442e-b8ce-231bcbbb434e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.233676 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b62eafcb-c14a-442e-b8ce-231bcbbb434e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b62eafcb-c14a-442e-b8ce-231bcbbb434e" (UID: "b62eafcb-c14a-442e-b8ce-231bcbbb434e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.304131 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv5t4\" (UniqueName: \"kubernetes.io/projected/b62eafcb-c14a-442e-b8ce-231bcbbb434e-kube-api-access-bv5t4\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.304175 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b62eafcb-c14a-442e-b8ce-231bcbbb434e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.304192 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b62eafcb-c14a-442e-b8ce-231bcbbb434e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.304204 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b62eafcb-c14a-442e-b8ce-231bcbbb434e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.877712 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b62eafcb-c14a-442e-b8ce-231bcbbb434e","Type":"ContainerDied","Data":"88b843e5d937b83f6e16fd08a6d47a1e888d7e9a13eb4c82c9744ab21cf54ff2"} Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.877770 4808 scope.go:117] "RemoveContainer" containerID="c74d6e32d7fe951e38a94f91562a264d2be51496678b451af2e176ad4583ffe0" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.877772 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.907468 4808 scope.go:117] "RemoveContainer" containerID="b114fd29e32611d2048b0bd3384298dc3a6e8bfd6fc5905cb04b5627cfa39155" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.914569 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.922631 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.946137 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:45:23 crc kubenswrapper[4808]: E1124 17:45:23.946545 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b62eafcb-c14a-442e-b8ce-231bcbbb434e" containerName="nova-metadata-metadata" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.946566 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b62eafcb-c14a-442e-b8ce-231bcbbb434e" containerName="nova-metadata-metadata" Nov 24 17:45:23 crc kubenswrapper[4808]: E1124 17:45:23.946592 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b62eafcb-c14a-442e-b8ce-231bcbbb434e" containerName="nova-metadata-log" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.946599 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b62eafcb-c14a-442e-b8ce-231bcbbb434e" containerName="nova-metadata-log" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.946789 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b62eafcb-c14a-442e-b8ce-231bcbbb434e" containerName="nova-metadata-log" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.946809 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b62eafcb-c14a-442e-b8ce-231bcbbb434e" containerName="nova-metadata-metadata" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.947806 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.950532 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.951462 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 17:45:23 crc kubenswrapper[4808]: I1124 17:45:23.954432 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.118572 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.118625 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7kjl\" (UniqueName: \"kubernetes.io/projected/7603ec6b-d363-485f-bbb5-d1de3d6df76e-kube-api-access-s7kjl\") pod \"nova-metadata-0\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.118667 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7603ec6b-d363-485f-bbb5-d1de3d6df76e-logs\") pod \"nova-metadata-0\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.118716 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.118805 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-config-data\") pod \"nova-metadata-0\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.221130 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.221258 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-config-data\") pod \"nova-metadata-0\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.221308 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.221340 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7kjl\" (UniqueName: \"kubernetes.io/projected/7603ec6b-d363-485f-bbb5-d1de3d6df76e-kube-api-access-s7kjl\") pod \"nova-metadata-0\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.221377 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7603ec6b-d363-485f-bbb5-d1de3d6df76e-logs\") pod \"nova-metadata-0\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.222697 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7603ec6b-d363-485f-bbb5-d1de3d6df76e-logs\") pod \"nova-metadata-0\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.228994 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.230457 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-config-data\") pod \"nova-metadata-0\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.230457 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.249662 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7kjl\" (UniqueName: \"kubernetes.io/projected/7603ec6b-d363-485f-bbb5-d1de3d6df76e-kube-api-access-s7kjl\") pod \"nova-metadata-0\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.273743 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.369815 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b62eafcb-c14a-442e-b8ce-231bcbbb434e" path="/var/lib/kubelet/pods/b62eafcb-c14a-442e-b8ce-231bcbbb434e/volumes" Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.719362 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:45:24 crc kubenswrapper[4808]: I1124 17:45:24.888372 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7603ec6b-d363-485f-bbb5-d1de3d6df76e","Type":"ContainerStarted","Data":"8f17407c3d133caf0836b38e6bea095178b105d861edfe32fbecbe1e10b4cef8"} Nov 24 17:45:25 crc kubenswrapper[4808]: I1124 17:45:25.871235 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 17:45:25 crc kubenswrapper[4808]: I1124 17:45:25.871294 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 17:45:25 crc kubenswrapper[4808]: I1124 17:45:25.899847 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7603ec6b-d363-485f-bbb5-d1de3d6df76e","Type":"ContainerStarted","Data":"7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f"} Nov 24 17:45:25 crc kubenswrapper[4808]: I1124 17:45:25.899909 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7603ec6b-d363-485f-bbb5-d1de3d6df76e","Type":"ContainerStarted","Data":"8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41"} Nov 24 17:45:25 crc kubenswrapper[4808]: I1124 17:45:25.903125 4808 generic.go:334] "Generic (PLEG): container finished" podID="1ff6c89f-2920-4894-8036-bbf316ecd48b" containerID="522929826dc3bfc36bdc89b2ee5bc6c4d9d79ae11afa1c70662522d45d6b603d" exitCode=0 Nov 24 17:45:25 crc kubenswrapper[4808]: I1124 17:45:25.903165 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lfpxw" event={"ID":"1ff6c89f-2920-4894-8036-bbf316ecd48b","Type":"ContainerDied","Data":"522929826dc3bfc36bdc89b2ee5bc6c4d9d79ae11afa1c70662522d45d6b603d"} Nov 24 17:45:25 crc kubenswrapper[4808]: I1124 17:45:25.910914 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 17:45:25 crc kubenswrapper[4808]: I1124 17:45:25.910949 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 17:45:25 crc kubenswrapper[4808]: I1124 17:45:25.926524 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.92649924 podStartE2EDuration="2.92649924s" podCreationTimestamp="2025-11-24 17:45:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:45:25.923210066 +0000 UTC m=+1118.520877868" watchObservedRunningTime="2025-11-24 17:45:25.92649924 +0000 UTC m=+1118.524167052" Nov 24 17:45:25 crc kubenswrapper[4808]: I1124 17:45:25.937254 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 17:45:25 crc kubenswrapper[4808]: I1124 17:45:25.980087 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.051217 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.298145 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.385526 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-j2p9p"] Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.386117 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" podUID="d02a83ff-86ec-4de5-b9fb-47d5082c387b" containerName="dnsmasq-dns" containerID="cri-o://ee6fe4ea8c6951db3a9fe5c4b758165bf16ef1a9d8063f6b9028ec870d332b0a" gracePeriod=10 Nov 24 17:45:26 crc kubenswrapper[4808]: E1124 17:45:26.657651 4808 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd02a83ff_86ec_4de5_b9fb_47d5082c387b.slice/crio-conmon-ee6fe4ea8c6951db3a9fe5c4b758165bf16ef1a9d8063f6b9028ec870d332b0a.scope\": RecentStats: unable to find data in memory cache]" Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.906325 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.913806 4808 generic.go:334] "Generic (PLEG): container finished" podID="d02a83ff-86ec-4de5-b9fb-47d5082c387b" containerID="ee6fe4ea8c6951db3a9fe5c4b758165bf16ef1a9d8063f6b9028ec870d332b0a" exitCode=0 Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.913862 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" event={"ID":"d02a83ff-86ec-4de5-b9fb-47d5082c387b","Type":"ContainerDied","Data":"ee6fe4ea8c6951db3a9fe5c4b758165bf16ef1a9d8063f6b9028ec870d332b0a"} Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.913921 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" event={"ID":"d02a83ff-86ec-4de5-b9fb-47d5082c387b","Type":"ContainerDied","Data":"70a5da4417edd76c05b1382f11b01d3808d02e8cf2c39f404919e98934110a91"} Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.913938 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-j2p9p" Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.913945 4808 scope.go:117] "RemoveContainer" containerID="ee6fe4ea8c6951db3a9fe5c4b758165bf16ef1a9d8063f6b9028ec870d332b0a" Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.952175 4808 scope.go:117] "RemoveContainer" containerID="34652f239e549030e3f689284db06659b83af24b7723483a27e6810f27d80f00" Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.971094 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-dns-svc\") pod \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.971203 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-dns-swift-storage-0\") pod \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.971247 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-ovsdbserver-nb\") pod \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.971290 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-ovsdbserver-sb\") pod \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.971372 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-config\") pod \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.971423 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gp8l\" (UniqueName: \"kubernetes.io/projected/d02a83ff-86ec-4de5-b9fb-47d5082c387b-kube-api-access-9gp8l\") pod \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\" (UID: \"d02a83ff-86ec-4de5-b9fb-47d5082c387b\") " Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.989473 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d02a83ff-86ec-4de5-b9fb-47d5082c387b-kube-api-access-9gp8l" (OuterVolumeSpecName: "kube-api-access-9gp8l") pod "d02a83ff-86ec-4de5-b9fb-47d5082c387b" (UID: "d02a83ff-86ec-4de5-b9fb-47d5082c387b"). InnerVolumeSpecName "kube-api-access-9gp8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.994231 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 17:45:26 crc kubenswrapper[4808]: I1124 17:45:26.994489 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.016580 4808 scope.go:117] "RemoveContainer" containerID="ee6fe4ea8c6951db3a9fe5c4b758165bf16ef1a9d8063f6b9028ec870d332b0a" Nov 24 17:45:27 crc kubenswrapper[4808]: E1124 17:45:27.018143 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee6fe4ea8c6951db3a9fe5c4b758165bf16ef1a9d8063f6b9028ec870d332b0a\": container with ID starting with ee6fe4ea8c6951db3a9fe5c4b758165bf16ef1a9d8063f6b9028ec870d332b0a not found: ID does not exist" containerID="ee6fe4ea8c6951db3a9fe5c4b758165bf16ef1a9d8063f6b9028ec870d332b0a" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.018186 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee6fe4ea8c6951db3a9fe5c4b758165bf16ef1a9d8063f6b9028ec870d332b0a"} err="failed to get container status \"ee6fe4ea8c6951db3a9fe5c4b758165bf16ef1a9d8063f6b9028ec870d332b0a\": rpc error: code = NotFound desc = could not find container \"ee6fe4ea8c6951db3a9fe5c4b758165bf16ef1a9d8063f6b9028ec870d332b0a\": container with ID starting with ee6fe4ea8c6951db3a9fe5c4b758165bf16ef1a9d8063f6b9028ec870d332b0a not found: ID does not exist" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.018210 4808 scope.go:117] "RemoveContainer" containerID="34652f239e549030e3f689284db06659b83af24b7723483a27e6810f27d80f00" Nov 24 17:45:27 crc kubenswrapper[4808]: E1124 17:45:27.020483 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34652f239e549030e3f689284db06659b83af24b7723483a27e6810f27d80f00\": container with ID starting with 34652f239e549030e3f689284db06659b83af24b7723483a27e6810f27d80f00 not found: ID does not exist" containerID="34652f239e549030e3f689284db06659b83af24b7723483a27e6810f27d80f00" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.020822 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34652f239e549030e3f689284db06659b83af24b7723483a27e6810f27d80f00"} err="failed to get container status \"34652f239e549030e3f689284db06659b83af24b7723483a27e6810f27d80f00\": rpc error: code = NotFound desc = could not find container \"34652f239e549030e3f689284db06659b83af24b7723483a27e6810f27d80f00\": container with ID starting with 34652f239e549030e3f689284db06659b83af24b7723483a27e6810f27d80f00 not found: ID does not exist" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.036052 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d02a83ff-86ec-4de5-b9fb-47d5082c387b" (UID: "d02a83ff-86ec-4de5-b9fb-47d5082c387b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.061083 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d02a83ff-86ec-4de5-b9fb-47d5082c387b" (UID: "d02a83ff-86ec-4de5-b9fb-47d5082c387b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.069802 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d02a83ff-86ec-4de5-b9fb-47d5082c387b" (UID: "d02a83ff-86ec-4de5-b9fb-47d5082c387b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.070718 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-config" (OuterVolumeSpecName: "config") pod "d02a83ff-86ec-4de5-b9fb-47d5082c387b" (UID: "d02a83ff-86ec-4de5-b9fb-47d5082c387b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.074075 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gp8l\" (UniqueName: \"kubernetes.io/projected/d02a83ff-86ec-4de5-b9fb-47d5082c387b-kube-api-access-9gp8l\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.074104 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.074118 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.074131 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.074145 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.099564 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d02a83ff-86ec-4de5-b9fb-47d5082c387b" (UID: "d02a83ff-86ec-4de5-b9fb-47d5082c387b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.175492 4808 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d02a83ff-86ec-4de5-b9fb-47d5082c387b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.254492 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-j2p9p"] Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.261569 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.264346 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-j2p9p"] Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.276283 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-combined-ca-bundle\") pod \"1ff6c89f-2920-4894-8036-bbf316ecd48b\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.276498 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-config-data\") pod \"1ff6c89f-2920-4894-8036-bbf316ecd48b\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.276530 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28s6k\" (UniqueName: \"kubernetes.io/projected/1ff6c89f-2920-4894-8036-bbf316ecd48b-kube-api-access-28s6k\") pod \"1ff6c89f-2920-4894-8036-bbf316ecd48b\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.276606 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-scripts\") pod \"1ff6c89f-2920-4894-8036-bbf316ecd48b\" (UID: \"1ff6c89f-2920-4894-8036-bbf316ecd48b\") " Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.280359 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-scripts" (OuterVolumeSpecName: "scripts") pod "1ff6c89f-2920-4894-8036-bbf316ecd48b" (UID: "1ff6c89f-2920-4894-8036-bbf316ecd48b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.282008 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ff6c89f-2920-4894-8036-bbf316ecd48b-kube-api-access-28s6k" (OuterVolumeSpecName: "kube-api-access-28s6k") pod "1ff6c89f-2920-4894-8036-bbf316ecd48b" (UID: "1ff6c89f-2920-4894-8036-bbf316ecd48b"). InnerVolumeSpecName "kube-api-access-28s6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.303614 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-config-data" (OuterVolumeSpecName: "config-data") pod "1ff6c89f-2920-4894-8036-bbf316ecd48b" (UID: "1ff6c89f-2920-4894-8036-bbf316ecd48b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.332199 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ff6c89f-2920-4894-8036-bbf316ecd48b" (UID: "1ff6c89f-2920-4894-8036-bbf316ecd48b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.379374 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.379411 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.379422 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28s6k\" (UniqueName: \"kubernetes.io/projected/1ff6c89f-2920-4894-8036-bbf316ecd48b-kube-api-access-28s6k\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.379430 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ff6c89f-2920-4894-8036-bbf316ecd48b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.927534 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lfpxw" event={"ID":"1ff6c89f-2920-4894-8036-bbf316ecd48b","Type":"ContainerDied","Data":"3e4783e062ee65c957d20483bcaba76a6f8872edfd2407c8709d0619bd1dc770"} Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.927581 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e4783e062ee65c957d20483bcaba76a6f8872edfd2407c8709d0619bd1dc770" Nov 24 17:45:27 crc kubenswrapper[4808]: I1124 17:45:27.927934 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lfpxw" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.146140 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.146399 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" containerName="nova-api-log" containerID="cri-o://07ce607292a080bb811b46df930b8b0039a46023a74a45bff26944dc5fa5c773" gracePeriod=30 Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.146537 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" containerName="nova-api-api" containerID="cri-o://84f7d4e24d959ce8f11afb76b8220e2fb780828e8701136618a2be208a1d5dfa" gracePeriod=30 Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.155342 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.155537 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8647155b-5589-4daa-b1a8-5bbffcc6e3f4" containerName="nova-scheduler-scheduler" containerID="cri-o://f0fa9927bcfe60084ac3154126f4def663f8221c09275f1c8bdcb7c2519764d6" gracePeriod=30 Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.193114 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.193412 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7603ec6b-d363-485f-bbb5-d1de3d6df76e" containerName="nova-metadata-log" containerID="cri-o://8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41" gracePeriod=30 Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.193517 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7603ec6b-d363-485f-bbb5-d1de3d6df76e" containerName="nova-metadata-metadata" containerID="cri-o://7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f" gracePeriod=30 Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.357263 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d02a83ff-86ec-4de5-b9fb-47d5082c387b" path="/var/lib/kubelet/pods/d02a83ff-86ec-4de5-b9fb-47d5082c387b/volumes" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.792795 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.813031 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7kjl\" (UniqueName: \"kubernetes.io/projected/7603ec6b-d363-485f-bbb5-d1de3d6df76e-kube-api-access-s7kjl\") pod \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.813135 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-nova-metadata-tls-certs\") pod \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.813212 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-config-data\") pod \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.813282 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-combined-ca-bundle\") pod \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.813304 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7603ec6b-d363-485f-bbb5-d1de3d6df76e-logs\") pod \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\" (UID: \"7603ec6b-d363-485f-bbb5-d1de3d6df76e\") " Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.814513 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7603ec6b-d363-485f-bbb5-d1de3d6df76e-logs" (OuterVolumeSpecName: "logs") pod "7603ec6b-d363-485f-bbb5-d1de3d6df76e" (UID: "7603ec6b-d363-485f-bbb5-d1de3d6df76e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.820119 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7603ec6b-d363-485f-bbb5-d1de3d6df76e-kube-api-access-s7kjl" (OuterVolumeSpecName: "kube-api-access-s7kjl") pod "7603ec6b-d363-485f-bbb5-d1de3d6df76e" (UID: "7603ec6b-d363-485f-bbb5-d1de3d6df76e"). InnerVolumeSpecName "kube-api-access-s7kjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.848326 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-config-data" (OuterVolumeSpecName: "config-data") pod "7603ec6b-d363-485f-bbb5-d1de3d6df76e" (UID: "7603ec6b-d363-485f-bbb5-d1de3d6df76e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.871587 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7603ec6b-d363-485f-bbb5-d1de3d6df76e" (UID: "7603ec6b-d363-485f-bbb5-d1de3d6df76e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.884245 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "7603ec6b-d363-485f-bbb5-d1de3d6df76e" (UID: "7603ec6b-d363-485f-bbb5-d1de3d6df76e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.915275 4808 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.915308 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.915320 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7603ec6b-d363-485f-bbb5-d1de3d6df76e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.915328 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7603ec6b-d363-485f-bbb5-d1de3d6df76e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.915338 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7kjl\" (UniqueName: \"kubernetes.io/projected/7603ec6b-d363-485f-bbb5-d1de3d6df76e-kube-api-access-s7kjl\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.942574 4808 generic.go:334] "Generic (PLEG): container finished" podID="1373d3ee-b91b-413e-af94-40d6c55c1917" containerID="82ba366ba8d51289ea732089d13e7cadb0dd0c2c292e82a6b1e301f97a7ae5ab" exitCode=0 Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.942634 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rfs5s" event={"ID":"1373d3ee-b91b-413e-af94-40d6c55c1917","Type":"ContainerDied","Data":"82ba366ba8d51289ea732089d13e7cadb0dd0c2c292e82a6b1e301f97a7ae5ab"} Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.946834 4808 generic.go:334] "Generic (PLEG): container finished" podID="7603ec6b-d363-485f-bbb5-d1de3d6df76e" containerID="7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f" exitCode=0 Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.947180 4808 generic.go:334] "Generic (PLEG): container finished" podID="7603ec6b-d363-485f-bbb5-d1de3d6df76e" containerID="8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41" exitCode=143 Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.947073 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7603ec6b-d363-485f-bbb5-d1de3d6df76e","Type":"ContainerDied","Data":"7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f"} Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.947264 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7603ec6b-d363-485f-bbb5-d1de3d6df76e","Type":"ContainerDied","Data":"8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41"} Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.947283 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7603ec6b-d363-485f-bbb5-d1de3d6df76e","Type":"ContainerDied","Data":"8f17407c3d133caf0836b38e6bea095178b105d861edfe32fbecbe1e10b4cef8"} Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.947303 4808 scope.go:117] "RemoveContainer" containerID="7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.947112 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.950191 4808 generic.go:334] "Generic (PLEG): container finished" podID="9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" containerID="07ce607292a080bb811b46df930b8b0039a46023a74a45bff26944dc5fa5c773" exitCode=143 Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.950227 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599","Type":"ContainerDied","Data":"07ce607292a080bb811b46df930b8b0039a46023a74a45bff26944dc5fa5c773"} Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.987745 4808 scope.go:117] "RemoveContainer" containerID="8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41" Nov 24 17:45:28 crc kubenswrapper[4808]: I1124 17:45:28.998445 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.019487 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.021587 4808 scope.go:117] "RemoveContainer" containerID="7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f" Nov 24 17:45:29 crc kubenswrapper[4808]: E1124 17:45:29.022086 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f\": container with ID starting with 7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f not found: ID does not exist" containerID="7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.022118 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f"} err="failed to get container status \"7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f\": rpc error: code = NotFound desc = could not find container \"7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f\": container with ID starting with 7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f not found: ID does not exist" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.022138 4808 scope.go:117] "RemoveContainer" containerID="8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41" Nov 24 17:45:29 crc kubenswrapper[4808]: E1124 17:45:29.022720 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41\": container with ID starting with 8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41 not found: ID does not exist" containerID="8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.022755 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41"} err="failed to get container status \"8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41\": rpc error: code = NotFound desc = could not find container \"8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41\": container with ID starting with 8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41 not found: ID does not exist" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.022772 4808 scope.go:117] "RemoveContainer" containerID="7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.023259 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f"} err="failed to get container status \"7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f\": rpc error: code = NotFound desc = could not find container \"7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f\": container with ID starting with 7e88c5a6296759ba70c999ea1ff855b7123639bed2cd345c94e304706a6de76f not found: ID does not exist" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.023281 4808 scope.go:117] "RemoveContainer" containerID="8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.023599 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41"} err="failed to get container status \"8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41\": rpc error: code = NotFound desc = could not find container \"8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41\": container with ID starting with 8a3926eb296dfd5c2bfe3bc18ef2116c1313b35ef1f21b00ec31b8bb24a1bf41 not found: ID does not exist" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.029782 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:45:29 crc kubenswrapper[4808]: E1124 17:45:29.030238 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7603ec6b-d363-485f-bbb5-d1de3d6df76e" containerName="nova-metadata-log" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.030251 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7603ec6b-d363-485f-bbb5-d1de3d6df76e" containerName="nova-metadata-log" Nov 24 17:45:29 crc kubenswrapper[4808]: E1124 17:45:29.030263 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ff6c89f-2920-4894-8036-bbf316ecd48b" containerName="nova-manage" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.030269 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ff6c89f-2920-4894-8036-bbf316ecd48b" containerName="nova-manage" Nov 24 17:45:29 crc kubenswrapper[4808]: E1124 17:45:29.030280 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7603ec6b-d363-485f-bbb5-d1de3d6df76e" containerName="nova-metadata-metadata" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.030286 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7603ec6b-d363-485f-bbb5-d1de3d6df76e" containerName="nova-metadata-metadata" Nov 24 17:45:29 crc kubenswrapper[4808]: E1124 17:45:29.030299 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d02a83ff-86ec-4de5-b9fb-47d5082c387b" containerName="dnsmasq-dns" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.030304 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d02a83ff-86ec-4de5-b9fb-47d5082c387b" containerName="dnsmasq-dns" Nov 24 17:45:29 crc kubenswrapper[4808]: E1124 17:45:29.030328 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d02a83ff-86ec-4de5-b9fb-47d5082c387b" containerName="init" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.030334 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d02a83ff-86ec-4de5-b9fb-47d5082c387b" containerName="init" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.030501 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="d02a83ff-86ec-4de5-b9fb-47d5082c387b" containerName="dnsmasq-dns" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.030516 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7603ec6b-d363-485f-bbb5-d1de3d6df76e" containerName="nova-metadata-metadata" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.030527 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ff6c89f-2920-4894-8036-bbf316ecd48b" containerName="nova-manage" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.030536 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7603ec6b-d363-485f-bbb5-d1de3d6df76e" containerName="nova-metadata-log" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.031531 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.035960 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.036408 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.039057 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.118338 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-config-data\") pod \"nova-metadata-0\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.118441 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.118472 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6359988a-9887-4f69-a3ab-65de38bb758f-logs\") pod \"nova-metadata-0\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.118551 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.118587 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsdtm\" (UniqueName: \"kubernetes.io/projected/6359988a-9887-4f69-a3ab-65de38bb758f-kube-api-access-xsdtm\") pod \"nova-metadata-0\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.219798 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-config-data\") pod \"nova-metadata-0\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.220166 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.220289 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6359988a-9887-4f69-a3ab-65de38bb758f-logs\") pod \"nova-metadata-0\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.220418 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.220520 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsdtm\" (UniqueName: \"kubernetes.io/projected/6359988a-9887-4f69-a3ab-65de38bb758f-kube-api-access-xsdtm\") pod \"nova-metadata-0\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.221162 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6359988a-9887-4f69-a3ab-65de38bb758f-logs\") pod \"nova-metadata-0\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.223280 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-config-data\") pod \"nova-metadata-0\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.224250 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.225638 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.237238 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsdtm\" (UniqueName: \"kubernetes.io/projected/6359988a-9887-4f69-a3ab-65de38bb758f-kube-api-access-xsdtm\") pod \"nova-metadata-0\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.355725 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.781972 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:45:29 crc kubenswrapper[4808]: W1124 17:45:29.785634 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6359988a_9887_4f69_a3ab_65de38bb758f.slice/crio-8f48392722cb1b921279be478b98fe5092bfb19dd190473ae40e05fd58d27628 WatchSource:0}: Error finding container 8f48392722cb1b921279be478b98fe5092bfb19dd190473ae40e05fd58d27628: Status 404 returned error can't find the container with id 8f48392722cb1b921279be478b98fe5092bfb19dd190473ae40e05fd58d27628 Nov 24 17:45:29 crc kubenswrapper[4808]: I1124 17:45:29.973874 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6359988a-9887-4f69-a3ab-65de38bb758f","Type":"ContainerStarted","Data":"8f48392722cb1b921279be478b98fe5092bfb19dd190473ae40e05fd58d27628"} Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.237367 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.358230 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qnz9\" (UniqueName: \"kubernetes.io/projected/1373d3ee-b91b-413e-af94-40d6c55c1917-kube-api-access-8qnz9\") pod \"1373d3ee-b91b-413e-af94-40d6c55c1917\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.358344 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-combined-ca-bundle\") pod \"1373d3ee-b91b-413e-af94-40d6c55c1917\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.358406 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-scripts\") pod \"1373d3ee-b91b-413e-af94-40d6c55c1917\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.358448 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-config-data\") pod \"1373d3ee-b91b-413e-af94-40d6c55c1917\" (UID: \"1373d3ee-b91b-413e-af94-40d6c55c1917\") " Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.362285 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1373d3ee-b91b-413e-af94-40d6c55c1917-kube-api-access-8qnz9" (OuterVolumeSpecName: "kube-api-access-8qnz9") pod "1373d3ee-b91b-413e-af94-40d6c55c1917" (UID: "1373d3ee-b91b-413e-af94-40d6c55c1917"). InnerVolumeSpecName "kube-api-access-8qnz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.362038 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7603ec6b-d363-485f-bbb5-d1de3d6df76e" path="/var/lib/kubelet/pods/7603ec6b-d363-485f-bbb5-d1de3d6df76e/volumes" Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.379948 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-scripts" (OuterVolumeSpecName: "scripts") pod "1373d3ee-b91b-413e-af94-40d6c55c1917" (UID: "1373d3ee-b91b-413e-af94-40d6c55c1917"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.385129 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1373d3ee-b91b-413e-af94-40d6c55c1917" (UID: "1373d3ee-b91b-413e-af94-40d6c55c1917"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.403797 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-config-data" (OuterVolumeSpecName: "config-data") pod "1373d3ee-b91b-413e-af94-40d6c55c1917" (UID: "1373d3ee-b91b-413e-af94-40d6c55c1917"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.461664 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.461722 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.461736 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1373d3ee-b91b-413e-af94-40d6c55c1917-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.461749 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qnz9\" (UniqueName: \"kubernetes.io/projected/1373d3ee-b91b-413e-af94-40d6c55c1917-kube-api-access-8qnz9\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:30 crc kubenswrapper[4808]: E1124 17:45:30.873394 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f0fa9927bcfe60084ac3154126f4def663f8221c09275f1c8bdcb7c2519764d6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 17:45:30 crc kubenswrapper[4808]: E1124 17:45:30.875325 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f0fa9927bcfe60084ac3154126f4def663f8221c09275f1c8bdcb7c2519764d6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 17:45:30 crc kubenswrapper[4808]: E1124 17:45:30.877095 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f0fa9927bcfe60084ac3154126f4def663f8221c09275f1c8bdcb7c2519764d6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 17:45:30 crc kubenswrapper[4808]: E1124 17:45:30.877142 4808 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8647155b-5589-4daa-b1a8-5bbffcc6e3f4" containerName="nova-scheduler-scheduler" Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.987482 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6359988a-9887-4f69-a3ab-65de38bb758f","Type":"ContainerStarted","Data":"b4db06dd8a5501098c6b11178ae48cf956c33b8e9000b870120d429993d7f2c8"} Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.987535 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6359988a-9887-4f69-a3ab-65de38bb758f","Type":"ContainerStarted","Data":"3362edc7aa66c6fa8a37f7b7b785865f30c8b2e749f8de3faf42fd870b6616d7"} Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.993982 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rfs5s" event={"ID":"1373d3ee-b91b-413e-af94-40d6c55c1917","Type":"ContainerDied","Data":"661df439ac3ccacbc182cf9ad6947dee5f55f1e84595515e43259fd592a64257"} Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.994044 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="661df439ac3ccacbc182cf9ad6947dee5f55f1e84595515e43259fd592a64257" Nov 24 17:45:30 crc kubenswrapper[4808]: I1124 17:45:30.994148 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rfs5s" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.021617 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.021592202 podStartE2EDuration="3.021592202s" podCreationTimestamp="2025-11-24 17:45:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:45:31.017564527 +0000 UTC m=+1123.615232379" watchObservedRunningTime="2025-11-24 17:45:31.021592202 +0000 UTC m=+1123.619260004" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.044971 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 17:45:31 crc kubenswrapper[4808]: E1124 17:45:31.045660 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1373d3ee-b91b-413e-af94-40d6c55c1917" containerName="nova-cell1-conductor-db-sync" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.045695 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="1373d3ee-b91b-413e-af94-40d6c55c1917" containerName="nova-cell1-conductor-db-sync" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.046002 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="1373d3ee-b91b-413e-af94-40d6c55c1917" containerName="nova-cell1-conductor-db-sync" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.046893 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.049470 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.053904 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.175029 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7074b2db-ae2b-4fee-a890-6cc64cd46cf8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7074b2db-ae2b-4fee-a890-6cc64cd46cf8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.175182 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7074b2db-ae2b-4fee-a890-6cc64cd46cf8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7074b2db-ae2b-4fee-a890-6cc64cd46cf8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.175265 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsr2g\" (UniqueName: \"kubernetes.io/projected/7074b2db-ae2b-4fee-a890-6cc64cd46cf8-kube-api-access-jsr2g\") pod \"nova-cell1-conductor-0\" (UID: \"7074b2db-ae2b-4fee-a890-6cc64cd46cf8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.276425 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7074b2db-ae2b-4fee-a890-6cc64cd46cf8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7074b2db-ae2b-4fee-a890-6cc64cd46cf8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.276521 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7074b2db-ae2b-4fee-a890-6cc64cd46cf8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7074b2db-ae2b-4fee-a890-6cc64cd46cf8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.276594 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsr2g\" (UniqueName: \"kubernetes.io/projected/7074b2db-ae2b-4fee-a890-6cc64cd46cf8-kube-api-access-jsr2g\") pod \"nova-cell1-conductor-0\" (UID: \"7074b2db-ae2b-4fee-a890-6cc64cd46cf8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.281521 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7074b2db-ae2b-4fee-a890-6cc64cd46cf8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7074b2db-ae2b-4fee-a890-6cc64cd46cf8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.290813 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7074b2db-ae2b-4fee-a890-6cc64cd46cf8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7074b2db-ae2b-4fee-a890-6cc64cd46cf8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.292920 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsr2g\" (UniqueName: \"kubernetes.io/projected/7074b2db-ae2b-4fee-a890-6cc64cd46cf8-kube-api-access-jsr2g\") pod \"nova-cell1-conductor-0\" (UID: \"7074b2db-ae2b-4fee-a890-6cc64cd46cf8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.373684 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 17:45:31 crc kubenswrapper[4808]: I1124 17:45:31.830330 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 17:45:31 crc kubenswrapper[4808]: W1124 17:45:31.834386 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7074b2db_ae2b_4fee_a890_6cc64cd46cf8.slice/crio-004bb55d7db4fd43586c0dbeaefd009da8064e7bc56751b760ca9964d43161b6 WatchSource:0}: Error finding container 004bb55d7db4fd43586c0dbeaefd009da8064e7bc56751b760ca9964d43161b6: Status 404 returned error can't find the container with id 004bb55d7db4fd43586c0dbeaefd009da8064e7bc56751b760ca9964d43161b6 Nov 24 17:45:32 crc kubenswrapper[4808]: I1124 17:45:32.003943 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7074b2db-ae2b-4fee-a890-6cc64cd46cf8","Type":"ContainerStarted","Data":"004bb55d7db4fd43586c0dbeaefd009da8064e7bc56751b760ca9964d43161b6"} Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.014839 4808 generic.go:334] "Generic (PLEG): container finished" podID="8647155b-5589-4daa-b1a8-5bbffcc6e3f4" containerID="f0fa9927bcfe60084ac3154126f4def663f8221c09275f1c8bdcb7c2519764d6" exitCode=0 Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.014926 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8647155b-5589-4daa-b1a8-5bbffcc6e3f4","Type":"ContainerDied","Data":"f0fa9927bcfe60084ac3154126f4def663f8221c09275f1c8bdcb7c2519764d6"} Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.018254 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7074b2db-ae2b-4fee-a890-6cc64cd46cf8","Type":"ContainerStarted","Data":"b7e1370e4aa6ef2d0c3c2556c7c6cfb9cf0b8e36e6dffa5677aa2a768b212a7b"} Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.018398 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.021114 4808 generic.go:334] "Generic (PLEG): container finished" podID="9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" containerID="84f7d4e24d959ce8f11afb76b8220e2fb780828e8701136618a2be208a1d5dfa" exitCode=0 Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.021176 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599","Type":"ContainerDied","Data":"84f7d4e24d959ce8f11afb76b8220e2fb780828e8701136618a2be208a1d5dfa"} Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.021209 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599","Type":"ContainerDied","Data":"484b416078ac71035dc34e3bae43ef0532fa8c7375d5dde9004f848103812867"} Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.021225 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="484b416078ac71035dc34e3bae43ef0532fa8c7375d5dde9004f848103812867" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.043390 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.043355553 podStartE2EDuration="2.043355553s" podCreationTimestamp="2025-11-24 17:45:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:45:33.040387519 +0000 UTC m=+1125.638055331" watchObservedRunningTime="2025-11-24 17:45:33.043355553 +0000 UTC m=+1125.641023355" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.089818 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.115091 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-config-data\") pod \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.115202 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjb8z\" (UniqueName: \"kubernetes.io/projected/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-kube-api-access-pjb8z\") pod \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.115931 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-combined-ca-bundle\") pod \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.116081 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-logs\") pod \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\" (UID: \"9912b2d7-4ac6-41f6-be1c-cbaf3cd63599\") " Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.117813 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-logs" (OuterVolumeSpecName: "logs") pod "9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" (UID: "9912b2d7-4ac6-41f6-be1c-cbaf3cd63599"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.118246 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.127870 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-kube-api-access-pjb8z" (OuterVolumeSpecName: "kube-api-access-pjb8z") pod "9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" (UID: "9912b2d7-4ac6-41f6-be1c-cbaf3cd63599"). InnerVolumeSpecName "kube-api-access-pjb8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.162929 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" (UID: "9912b2d7-4ac6-41f6-be1c-cbaf3cd63599"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.166990 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-config-data" (OuterVolumeSpecName: "config-data") pod "9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" (UID: "9912b2d7-4ac6-41f6-be1c-cbaf3cd63599"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.219661 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.219705 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjb8z\" (UniqueName: \"kubernetes.io/projected/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-kube-api-access-pjb8z\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.219714 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.313452 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.422466 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-combined-ca-bundle\") pod \"8647155b-5589-4daa-b1a8-5bbffcc6e3f4\" (UID: \"8647155b-5589-4daa-b1a8-5bbffcc6e3f4\") " Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.422512 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnkpz\" (UniqueName: \"kubernetes.io/projected/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-kube-api-access-bnkpz\") pod \"8647155b-5589-4daa-b1a8-5bbffcc6e3f4\" (UID: \"8647155b-5589-4daa-b1a8-5bbffcc6e3f4\") " Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.422655 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-config-data\") pod \"8647155b-5589-4daa-b1a8-5bbffcc6e3f4\" (UID: \"8647155b-5589-4daa-b1a8-5bbffcc6e3f4\") " Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.427988 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-kube-api-access-bnkpz" (OuterVolumeSpecName: "kube-api-access-bnkpz") pod "8647155b-5589-4daa-b1a8-5bbffcc6e3f4" (UID: "8647155b-5589-4daa-b1a8-5bbffcc6e3f4"). InnerVolumeSpecName "kube-api-access-bnkpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.449413 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8647155b-5589-4daa-b1a8-5bbffcc6e3f4" (UID: "8647155b-5589-4daa-b1a8-5bbffcc6e3f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.450937 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-config-data" (OuterVolumeSpecName: "config-data") pod "8647155b-5589-4daa-b1a8-5bbffcc6e3f4" (UID: "8647155b-5589-4daa-b1a8-5bbffcc6e3f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.524406 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.524443 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnkpz\" (UniqueName: \"kubernetes.io/projected/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-kube-api-access-bnkpz\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.524459 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8647155b-5589-4daa-b1a8-5bbffcc6e3f4-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:33 crc kubenswrapper[4808]: I1124 17:45:33.954553 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.040756 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.041117 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8647155b-5589-4daa-b1a8-5bbffcc6e3f4","Type":"ContainerDied","Data":"e9190293f873ef38dcf871c397b6dfe8aec3c4c3515abff32db6ec7a5a3daefe"} Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.041206 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.041263 4808 scope.go:117] "RemoveContainer" containerID="f0fa9927bcfe60084ac3154126f4def663f8221c09275f1c8bdcb7c2519764d6" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.258045 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.269334 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.288204 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:45:34 crc kubenswrapper[4808]: E1124 17:45:34.288633 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" containerName="nova-api-log" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.288654 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" containerName="nova-api-log" Nov 24 17:45:34 crc kubenswrapper[4808]: E1124 17:45:34.288680 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8647155b-5589-4daa-b1a8-5bbffcc6e3f4" containerName="nova-scheduler-scheduler" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.288689 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8647155b-5589-4daa-b1a8-5bbffcc6e3f4" containerName="nova-scheduler-scheduler" Nov 24 17:45:34 crc kubenswrapper[4808]: E1124 17:45:34.288736 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" containerName="nova-api-api" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.288744 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" containerName="nova-api-api" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.288982 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" containerName="nova-api-api" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.289004 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="8647155b-5589-4daa-b1a8-5bbffcc6e3f4" containerName="nova-scheduler-scheduler" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.289049 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" containerName="nova-api-log" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.301753 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.301888 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.307422 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.310982 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.326052 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.334572 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.336764 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.343748 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.346762 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.362711 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8647155b-5589-4daa-b1a8-5bbffcc6e3f4" path="/var/lib/kubelet/pods/8647155b-5589-4daa-b1a8-5bbffcc6e3f4/volumes" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.364168 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9912b2d7-4ac6-41f6-be1c-cbaf3cd63599" path="/var/lib/kubelet/pods/9912b2d7-4ac6-41f6-be1c-cbaf3cd63599/volumes" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.367504 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.367561 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.372965 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2d8bafdb-de51-40bc-b4d2-a4df903ebd38\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.373381 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-config-data\") pod \"nova-scheduler-0\" (UID: \"2d8bafdb-de51-40bc-b4d2-a4df903ebd38\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.373563 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwhhw\" (UniqueName: \"kubernetes.io/projected/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-kube-api-access-fwhhw\") pod \"nova-scheduler-0\" (UID: \"2d8bafdb-de51-40bc-b4d2-a4df903ebd38\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.475266 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2d8bafdb-de51-40bc-b4d2-a4df903ebd38\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.475369 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-config-data\") pod \"nova-scheduler-0\" (UID: \"2d8bafdb-de51-40bc-b4d2-a4df903ebd38\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.475452 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt6pl\" (UniqueName: \"kubernetes.io/projected/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-kube-api-access-lt6pl\") pod \"nova-api-0\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " pod="openstack/nova-api-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.475495 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " pod="openstack/nova-api-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.475538 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-logs\") pod \"nova-api-0\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " pod="openstack/nova-api-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.475587 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwhhw\" (UniqueName: \"kubernetes.io/projected/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-kube-api-access-fwhhw\") pod \"nova-scheduler-0\" (UID: \"2d8bafdb-de51-40bc-b4d2-a4df903ebd38\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.475682 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-config-data\") pod \"nova-api-0\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " pod="openstack/nova-api-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.480012 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2d8bafdb-de51-40bc-b4d2-a4df903ebd38\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.480108 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-config-data\") pod \"nova-scheduler-0\" (UID: \"2d8bafdb-de51-40bc-b4d2-a4df903ebd38\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.491175 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwhhw\" (UniqueName: \"kubernetes.io/projected/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-kube-api-access-fwhhw\") pod \"nova-scheduler-0\" (UID: \"2d8bafdb-de51-40bc-b4d2-a4df903ebd38\") " pod="openstack/nova-scheduler-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.578073 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt6pl\" (UniqueName: \"kubernetes.io/projected/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-kube-api-access-lt6pl\") pod \"nova-api-0\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " pod="openstack/nova-api-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.578139 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " pod="openstack/nova-api-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.578184 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-logs\") pod \"nova-api-0\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " pod="openstack/nova-api-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.578285 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-config-data\") pod \"nova-api-0\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " pod="openstack/nova-api-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.579337 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-logs\") pod \"nova-api-0\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " pod="openstack/nova-api-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.582752 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " pod="openstack/nova-api-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.583423 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-config-data\") pod \"nova-api-0\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " pod="openstack/nova-api-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.601933 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt6pl\" (UniqueName: \"kubernetes.io/projected/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-kube-api-access-lt6pl\") pod \"nova-api-0\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " pod="openstack/nova-api-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.629071 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:45:34 crc kubenswrapper[4808]: I1124 17:45:34.663081 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:45:35 crc kubenswrapper[4808]: I1124 17:45:35.114613 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:45:35 crc kubenswrapper[4808]: W1124 17:45:35.119400 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d8bafdb_de51_40bc_b4d2_a4df903ebd38.slice/crio-046369f04c0758ca2f68c3b3beb962b0ebc4b8107d4b7285e8c40e8ba7225f9f WatchSource:0}: Error finding container 046369f04c0758ca2f68c3b3beb962b0ebc4b8107d4b7285e8c40e8ba7225f9f: Status 404 returned error can't find the container with id 046369f04c0758ca2f68c3b3beb962b0ebc4b8107d4b7285e8c40e8ba7225f9f Nov 24 17:45:35 crc kubenswrapper[4808]: I1124 17:45:35.120906 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:45:35 crc kubenswrapper[4808]: W1124 17:45:35.121624 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd6137f4_f5fe_42e3_ae8d_7f25c746e7f2.slice/crio-27da97426c4ecd0919295d9cd1e0b3ce563520e248b768b3da42222a248b6162 WatchSource:0}: Error finding container 27da97426c4ecd0919295d9cd1e0b3ce563520e248b768b3da42222a248b6162: Status 404 returned error can't find the container with id 27da97426c4ecd0919295d9cd1e0b3ce563520e248b768b3da42222a248b6162 Nov 24 17:45:36 crc kubenswrapper[4808]: I1124 17:45:36.059383 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2d8bafdb-de51-40bc-b4d2-a4df903ebd38","Type":"ContainerStarted","Data":"7bff115e81d50a55a8c4cff31b951b1257483600b546b84121a9f4489a0f5ec4"} Nov 24 17:45:36 crc kubenswrapper[4808]: I1124 17:45:36.060070 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2d8bafdb-de51-40bc-b4d2-a4df903ebd38","Type":"ContainerStarted","Data":"046369f04c0758ca2f68c3b3beb962b0ebc4b8107d4b7285e8c40e8ba7225f9f"} Nov 24 17:45:36 crc kubenswrapper[4808]: I1124 17:45:36.062507 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2","Type":"ContainerStarted","Data":"2b04932f99541b6d928b2ee93d2391a5c7f583c935c474085c72c7f92e8e729e"} Nov 24 17:45:36 crc kubenswrapper[4808]: I1124 17:45:36.062542 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2","Type":"ContainerStarted","Data":"49b9f183139dfeb7ead25ef7550261b47c0cec64738cdd2156ad1699eb20efec"} Nov 24 17:45:36 crc kubenswrapper[4808]: I1124 17:45:36.062555 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2","Type":"ContainerStarted","Data":"27da97426c4ecd0919295d9cd1e0b3ce563520e248b768b3da42222a248b6162"} Nov 24 17:45:36 crc kubenswrapper[4808]: I1124 17:45:36.082825 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.082805618 podStartE2EDuration="2.082805618s" podCreationTimestamp="2025-11-24 17:45:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:45:36.076629032 +0000 UTC m=+1128.674296854" watchObservedRunningTime="2025-11-24 17:45:36.082805618 +0000 UTC m=+1128.680473420" Nov 24 17:45:36 crc kubenswrapper[4808]: I1124 17:45:36.100495 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.100476923 podStartE2EDuration="2.100476923s" podCreationTimestamp="2025-11-24 17:45:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:45:36.09057333 +0000 UTC m=+1128.688241132" watchObservedRunningTime="2025-11-24 17:45:36.100476923 +0000 UTC m=+1128.698144725" Nov 24 17:45:37 crc kubenswrapper[4808]: I1124 17:45:37.885514 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:45:37 crc kubenswrapper[4808]: I1124 17:45:37.886007 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d" containerName="kube-state-metrics" containerID="cri-o://ab233518ca094448e91b1460f50fe22092ce6bd95a14d785de74edc176ea346c" gracePeriod=30 Nov 24 17:45:38 crc kubenswrapper[4808]: I1124 17:45:38.081094 4808 generic.go:334] "Generic (PLEG): container finished" podID="f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d" containerID="ab233518ca094448e91b1460f50fe22092ce6bd95a14d785de74edc176ea346c" exitCode=2 Nov 24 17:45:38 crc kubenswrapper[4808]: I1124 17:45:38.081144 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d","Type":"ContainerDied","Data":"ab233518ca094448e91b1460f50fe22092ce6bd95a14d785de74edc176ea346c"} Nov 24 17:45:38 crc kubenswrapper[4808]: I1124 17:45:38.452936 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 17:45:38 crc kubenswrapper[4808]: I1124 17:45:38.554533 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px7c8\" (UniqueName: \"kubernetes.io/projected/f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d-kube-api-access-px7c8\") pod \"f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d\" (UID: \"f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d\") " Nov 24 17:45:38 crc kubenswrapper[4808]: I1124 17:45:38.562170 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d-kube-api-access-px7c8" (OuterVolumeSpecName: "kube-api-access-px7c8") pod "f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d" (UID: "f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d"). InnerVolumeSpecName "kube-api-access-px7c8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:45:38 crc kubenswrapper[4808]: I1124 17:45:38.656511 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px7c8\" (UniqueName: \"kubernetes.io/projected/f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d-kube-api-access-px7c8\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.089525 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d","Type":"ContainerDied","Data":"d036fd37de8907c34ba864b175d890f56cfb7841ad7d08a465d2a8c855ecffaa"} Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.089582 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.089597 4808 scope.go:117] "RemoveContainer" containerID="ab233518ca094448e91b1460f50fe22092ce6bd95a14d785de74edc176ea346c" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.135670 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.154360 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.165787 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:45:39 crc kubenswrapper[4808]: E1124 17:45:39.166184 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d" containerName="kube-state-metrics" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.166199 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d" containerName="kube-state-metrics" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.166435 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d" containerName="kube-state-metrics" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.167128 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.183727 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.184791 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.188052 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.267239 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/0c8560e5-eccf-482d-bc1b-74e44ccfeff1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"0c8560e5-eccf-482d-bc1b-74e44ccfeff1\") " pod="openstack/kube-state-metrics-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.267298 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8560e5-eccf-482d-bc1b-74e44ccfeff1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"0c8560e5-eccf-482d-bc1b-74e44ccfeff1\") " pod="openstack/kube-state-metrics-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.267367 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c8560e5-eccf-482d-bc1b-74e44ccfeff1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"0c8560e5-eccf-482d-bc1b-74e44ccfeff1\") " pod="openstack/kube-state-metrics-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.267429 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4p4w\" (UniqueName: \"kubernetes.io/projected/0c8560e5-eccf-482d-bc1b-74e44ccfeff1-kube-api-access-x4p4w\") pod \"kube-state-metrics-0\" (UID: \"0c8560e5-eccf-482d-bc1b-74e44ccfeff1\") " pod="openstack/kube-state-metrics-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.356060 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.356374 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.369567 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c8560e5-eccf-482d-bc1b-74e44ccfeff1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"0c8560e5-eccf-482d-bc1b-74e44ccfeff1\") " pod="openstack/kube-state-metrics-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.369657 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4p4w\" (UniqueName: \"kubernetes.io/projected/0c8560e5-eccf-482d-bc1b-74e44ccfeff1-kube-api-access-x4p4w\") pod \"kube-state-metrics-0\" (UID: \"0c8560e5-eccf-482d-bc1b-74e44ccfeff1\") " pod="openstack/kube-state-metrics-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.369786 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/0c8560e5-eccf-482d-bc1b-74e44ccfeff1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"0c8560e5-eccf-482d-bc1b-74e44ccfeff1\") " pod="openstack/kube-state-metrics-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.369819 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8560e5-eccf-482d-bc1b-74e44ccfeff1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"0c8560e5-eccf-482d-bc1b-74e44ccfeff1\") " pod="openstack/kube-state-metrics-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.375402 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8560e5-eccf-482d-bc1b-74e44ccfeff1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"0c8560e5-eccf-482d-bc1b-74e44ccfeff1\") " pod="openstack/kube-state-metrics-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.375637 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c8560e5-eccf-482d-bc1b-74e44ccfeff1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"0c8560e5-eccf-482d-bc1b-74e44ccfeff1\") " pod="openstack/kube-state-metrics-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.381565 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/0c8560e5-eccf-482d-bc1b-74e44ccfeff1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"0c8560e5-eccf-482d-bc1b-74e44ccfeff1\") " pod="openstack/kube-state-metrics-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.400934 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4p4w\" (UniqueName: \"kubernetes.io/projected/0c8560e5-eccf-482d-bc1b-74e44ccfeff1-kube-api-access-x4p4w\") pod \"kube-state-metrics-0\" (UID: \"0c8560e5-eccf-482d-bc1b-74e44ccfeff1\") " pod="openstack/kube-state-metrics-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.498008 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.630794 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.807510 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.807800 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="ceilometer-central-agent" containerID="cri-o://6b77fe1160b9842403f786da70a60e66398685601ab26dfdefd3d1c998398a64" gracePeriod=30 Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.808149 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="proxy-httpd" containerID="cri-o://f4ebc9ca0a99ca9f376a6f519e935c3ad22eb0a898b1d31ec7073c30f90aa108" gracePeriod=30 Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.808277 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="sg-core" containerID="cri-o://9e7e050c43538878ac97b5c541902c1729903db81dab306533c5bb4939ccb308" gracePeriod=30 Nov 24 17:45:39 crc kubenswrapper[4808]: I1124 17:45:39.808346 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="ceilometer-notification-agent" containerID="cri-o://1870350250d37b6b52d3bae0f6aa27541a7b6010f2c0de39dff35756d4a59b2c" gracePeriod=30 Nov 24 17:45:40 crc kubenswrapper[4808]: I1124 17:45:40.006530 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:45:40 crc kubenswrapper[4808]: W1124 17:45:40.006807 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c8560e5_eccf_482d_bc1b_74e44ccfeff1.slice/crio-e09e35c3559bb995327098f7b93c7137d3b0eabc3a442c2e473975aaac36d983 WatchSource:0}: Error finding container e09e35c3559bb995327098f7b93c7137d3b0eabc3a442c2e473975aaac36d983: Status 404 returned error can't find the container with id e09e35c3559bb995327098f7b93c7137d3b0eabc3a442c2e473975aaac36d983 Nov 24 17:45:40 crc kubenswrapper[4808]: I1124 17:45:40.103492 4808 generic.go:334] "Generic (PLEG): container finished" podID="77896689-86fa-4446-abd2-7c19ffde465c" containerID="f4ebc9ca0a99ca9f376a6f519e935c3ad22eb0a898b1d31ec7073c30f90aa108" exitCode=0 Nov 24 17:45:40 crc kubenswrapper[4808]: I1124 17:45:40.103542 4808 generic.go:334] "Generic (PLEG): container finished" podID="77896689-86fa-4446-abd2-7c19ffde465c" containerID="9e7e050c43538878ac97b5c541902c1729903db81dab306533c5bb4939ccb308" exitCode=2 Nov 24 17:45:40 crc kubenswrapper[4808]: I1124 17:45:40.103558 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77896689-86fa-4446-abd2-7c19ffde465c","Type":"ContainerDied","Data":"f4ebc9ca0a99ca9f376a6f519e935c3ad22eb0a898b1d31ec7073c30f90aa108"} Nov 24 17:45:40 crc kubenswrapper[4808]: I1124 17:45:40.103595 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77896689-86fa-4446-abd2-7c19ffde465c","Type":"ContainerDied","Data":"9e7e050c43538878ac97b5c541902c1729903db81dab306533c5bb4939ccb308"} Nov 24 17:45:40 crc kubenswrapper[4808]: I1124 17:45:40.106011 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0c8560e5-eccf-482d-bc1b-74e44ccfeff1","Type":"ContainerStarted","Data":"e09e35c3559bb995327098f7b93c7137d3b0eabc3a442c2e473975aaac36d983"} Nov 24 17:45:40 crc kubenswrapper[4808]: I1124 17:45:40.364861 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d" path="/var/lib/kubelet/pods/f3e65a2c-3b13-4bb1-8333-4ceae1f55d6d/volumes" Nov 24 17:45:40 crc kubenswrapper[4808]: I1124 17:45:40.373333 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6359988a-9887-4f69-a3ab-65de38bb758f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 17:45:40 crc kubenswrapper[4808]: I1124 17:45:40.373338 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6359988a-9887-4f69-a3ab-65de38bb758f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 17:45:41 crc kubenswrapper[4808]: I1124 17:45:41.124343 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0c8560e5-eccf-482d-bc1b-74e44ccfeff1","Type":"ContainerStarted","Data":"021599550bf6264c75e215acaa34e2a3e13212fd4327f1a0cb2b6e8eb7cc928f"} Nov 24 17:45:41 crc kubenswrapper[4808]: I1124 17:45:41.124505 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 17:45:41 crc kubenswrapper[4808]: I1124 17:45:41.127515 4808 generic.go:334] "Generic (PLEG): container finished" podID="77896689-86fa-4446-abd2-7c19ffde465c" containerID="6b77fe1160b9842403f786da70a60e66398685601ab26dfdefd3d1c998398a64" exitCode=0 Nov 24 17:45:41 crc kubenswrapper[4808]: I1124 17:45:41.127556 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77896689-86fa-4446-abd2-7c19ffde465c","Type":"ContainerDied","Data":"6b77fe1160b9842403f786da70a60e66398685601ab26dfdefd3d1c998398a64"} Nov 24 17:45:41 crc kubenswrapper[4808]: I1124 17:45:41.150763 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.770804246 podStartE2EDuration="2.150745136s" podCreationTimestamp="2025-11-24 17:45:39 +0000 UTC" firstStartedPulling="2025-11-24 17:45:40.011362336 +0000 UTC m=+1132.609030138" lastFinishedPulling="2025-11-24 17:45:40.391303226 +0000 UTC m=+1132.988971028" observedRunningTime="2025-11-24 17:45:41.144113656 +0000 UTC m=+1133.741781458" watchObservedRunningTime="2025-11-24 17:45:41.150745136 +0000 UTC m=+1133.748412928" Nov 24 17:45:41 crc kubenswrapper[4808]: I1124 17:45:41.404409 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.112278 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.148898 4808 generic.go:334] "Generic (PLEG): container finished" podID="77896689-86fa-4446-abd2-7c19ffde465c" containerID="1870350250d37b6b52d3bae0f6aa27541a7b6010f2c0de39dff35756d4a59b2c" exitCode=0 Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.148975 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.149039 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77896689-86fa-4446-abd2-7c19ffde465c","Type":"ContainerDied","Data":"1870350250d37b6b52d3bae0f6aa27541a7b6010f2c0de39dff35756d4a59b2c"} Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.149068 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77896689-86fa-4446-abd2-7c19ffde465c","Type":"ContainerDied","Data":"89ebf06e7247655950bc18445ce659362993bca483976296ca7340537fdf5877"} Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.149085 4808 scope.go:117] "RemoveContainer" containerID="f4ebc9ca0a99ca9f376a6f519e935c3ad22eb0a898b1d31ec7073c30f90aa108" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.170312 4808 scope.go:117] "RemoveContainer" containerID="9e7e050c43538878ac97b5c541902c1729903db81dab306533c5bb4939ccb308" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.191735 4808 scope.go:117] "RemoveContainer" containerID="1870350250d37b6b52d3bae0f6aa27541a7b6010f2c0de39dff35756d4a59b2c" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.212509 4808 scope.go:117] "RemoveContainer" containerID="6b77fe1160b9842403f786da70a60e66398685601ab26dfdefd3d1c998398a64" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.234777 4808 scope.go:117] "RemoveContainer" containerID="f4ebc9ca0a99ca9f376a6f519e935c3ad22eb0a898b1d31ec7073c30f90aa108" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.235263 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-config-data\") pod \"77896689-86fa-4446-abd2-7c19ffde465c\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.235327 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77896689-86fa-4446-abd2-7c19ffde465c-run-httpd\") pod \"77896689-86fa-4446-abd2-7c19ffde465c\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.235417 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-sg-core-conf-yaml\") pod \"77896689-86fa-4446-abd2-7c19ffde465c\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.235458 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77896689-86fa-4446-abd2-7c19ffde465c-log-httpd\") pod \"77896689-86fa-4446-abd2-7c19ffde465c\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.235497 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-scripts\") pod \"77896689-86fa-4446-abd2-7c19ffde465c\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.235533 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-combined-ca-bundle\") pod \"77896689-86fa-4446-abd2-7c19ffde465c\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.235661 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djr4w\" (UniqueName: \"kubernetes.io/projected/77896689-86fa-4446-abd2-7c19ffde465c-kube-api-access-djr4w\") pod \"77896689-86fa-4446-abd2-7c19ffde465c\" (UID: \"77896689-86fa-4446-abd2-7c19ffde465c\") " Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.236878 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77896689-86fa-4446-abd2-7c19ffde465c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "77896689-86fa-4446-abd2-7c19ffde465c" (UID: "77896689-86fa-4446-abd2-7c19ffde465c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:45:42 crc kubenswrapper[4808]: E1124 17:45:42.240129 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4ebc9ca0a99ca9f376a6f519e935c3ad22eb0a898b1d31ec7073c30f90aa108\": container with ID starting with f4ebc9ca0a99ca9f376a6f519e935c3ad22eb0a898b1d31ec7073c30f90aa108 not found: ID does not exist" containerID="f4ebc9ca0a99ca9f376a6f519e935c3ad22eb0a898b1d31ec7073c30f90aa108" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.240162 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4ebc9ca0a99ca9f376a6f519e935c3ad22eb0a898b1d31ec7073c30f90aa108"} err="failed to get container status \"f4ebc9ca0a99ca9f376a6f519e935c3ad22eb0a898b1d31ec7073c30f90aa108\": rpc error: code = NotFound desc = could not find container \"f4ebc9ca0a99ca9f376a6f519e935c3ad22eb0a898b1d31ec7073c30f90aa108\": container with ID starting with f4ebc9ca0a99ca9f376a6f519e935c3ad22eb0a898b1d31ec7073c30f90aa108 not found: ID does not exist" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.240189 4808 scope.go:117] "RemoveContainer" containerID="9e7e050c43538878ac97b5c541902c1729903db81dab306533c5bb4939ccb308" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.240496 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77896689-86fa-4446-abd2-7c19ffde465c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "77896689-86fa-4446-abd2-7c19ffde465c" (UID: "77896689-86fa-4446-abd2-7c19ffde465c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:45:42 crc kubenswrapper[4808]: E1124 17:45:42.242657 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e7e050c43538878ac97b5c541902c1729903db81dab306533c5bb4939ccb308\": container with ID starting with 9e7e050c43538878ac97b5c541902c1729903db81dab306533c5bb4939ccb308 not found: ID does not exist" containerID="9e7e050c43538878ac97b5c541902c1729903db81dab306533c5bb4939ccb308" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.242783 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e7e050c43538878ac97b5c541902c1729903db81dab306533c5bb4939ccb308"} err="failed to get container status \"9e7e050c43538878ac97b5c541902c1729903db81dab306533c5bb4939ccb308\": rpc error: code = NotFound desc = could not find container \"9e7e050c43538878ac97b5c541902c1729903db81dab306533c5bb4939ccb308\": container with ID starting with 9e7e050c43538878ac97b5c541902c1729903db81dab306533c5bb4939ccb308 not found: ID does not exist" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.242900 4808 scope.go:117] "RemoveContainer" containerID="1870350250d37b6b52d3bae0f6aa27541a7b6010f2c0de39dff35756d4a59b2c" Nov 24 17:45:42 crc kubenswrapper[4808]: E1124 17:45:42.243491 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1870350250d37b6b52d3bae0f6aa27541a7b6010f2c0de39dff35756d4a59b2c\": container with ID starting with 1870350250d37b6b52d3bae0f6aa27541a7b6010f2c0de39dff35756d4a59b2c not found: ID does not exist" containerID="1870350250d37b6b52d3bae0f6aa27541a7b6010f2c0de39dff35756d4a59b2c" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.243595 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1870350250d37b6b52d3bae0f6aa27541a7b6010f2c0de39dff35756d4a59b2c"} err="failed to get container status \"1870350250d37b6b52d3bae0f6aa27541a7b6010f2c0de39dff35756d4a59b2c\": rpc error: code = NotFound desc = could not find container \"1870350250d37b6b52d3bae0f6aa27541a7b6010f2c0de39dff35756d4a59b2c\": container with ID starting with 1870350250d37b6b52d3bae0f6aa27541a7b6010f2c0de39dff35756d4a59b2c not found: ID does not exist" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.243691 4808 scope.go:117] "RemoveContainer" containerID="6b77fe1160b9842403f786da70a60e66398685601ab26dfdefd3d1c998398a64" Nov 24 17:45:42 crc kubenswrapper[4808]: E1124 17:45:42.244454 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b77fe1160b9842403f786da70a60e66398685601ab26dfdefd3d1c998398a64\": container with ID starting with 6b77fe1160b9842403f786da70a60e66398685601ab26dfdefd3d1c998398a64 not found: ID does not exist" containerID="6b77fe1160b9842403f786da70a60e66398685601ab26dfdefd3d1c998398a64" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.244492 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b77fe1160b9842403f786da70a60e66398685601ab26dfdefd3d1c998398a64"} err="failed to get container status \"6b77fe1160b9842403f786da70a60e66398685601ab26dfdefd3d1c998398a64\": rpc error: code = NotFound desc = could not find container \"6b77fe1160b9842403f786da70a60e66398685601ab26dfdefd3d1c998398a64\": container with ID starting with 6b77fe1160b9842403f786da70a60e66398685601ab26dfdefd3d1c998398a64 not found: ID does not exist" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.258684 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77896689-86fa-4446-abd2-7c19ffde465c-kube-api-access-djr4w" (OuterVolumeSpecName: "kube-api-access-djr4w") pod "77896689-86fa-4446-abd2-7c19ffde465c" (UID: "77896689-86fa-4446-abd2-7c19ffde465c"). InnerVolumeSpecName "kube-api-access-djr4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.261847 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-scripts" (OuterVolumeSpecName: "scripts") pod "77896689-86fa-4446-abd2-7c19ffde465c" (UID: "77896689-86fa-4446-abd2-7c19ffde465c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.268330 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "77896689-86fa-4446-abd2-7c19ffde465c" (UID: "77896689-86fa-4446-abd2-7c19ffde465c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.318462 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77896689-86fa-4446-abd2-7c19ffde465c" (UID: "77896689-86fa-4446-abd2-7c19ffde465c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.337968 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djr4w\" (UniqueName: \"kubernetes.io/projected/77896689-86fa-4446-abd2-7c19ffde465c-kube-api-access-djr4w\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.338007 4808 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77896689-86fa-4446-abd2-7c19ffde465c-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.338033 4808 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.338045 4808 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77896689-86fa-4446-abd2-7c19ffde465c-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.338056 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.338066 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.349462 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-config-data" (OuterVolumeSpecName: "config-data") pod "77896689-86fa-4446-abd2-7c19ffde465c" (UID: "77896689-86fa-4446-abd2-7c19ffde465c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.439795 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77896689-86fa-4446-abd2-7c19ffde465c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.502695 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.518319 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.533124 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:45:42 crc kubenswrapper[4808]: E1124 17:45:42.533572 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="proxy-httpd" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.533593 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="proxy-httpd" Nov 24 17:45:42 crc kubenswrapper[4808]: E1124 17:45:42.533605 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="sg-core" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.533631 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="sg-core" Nov 24 17:45:42 crc kubenswrapper[4808]: E1124 17:45:42.533677 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="ceilometer-central-agent" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.533688 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="ceilometer-central-agent" Nov 24 17:45:42 crc kubenswrapper[4808]: E1124 17:45:42.533698 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="ceilometer-notification-agent" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.533706 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="ceilometer-notification-agent" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.533972 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="proxy-httpd" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.534008 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="ceilometer-notification-agent" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.534043 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="ceilometer-central-agent" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.534057 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="77896689-86fa-4446-abd2-7c19ffde465c" containerName="sg-core" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.536921 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.538456 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.541236 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.541994 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.547433 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.643494 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-scripts\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.643552 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msp87\" (UniqueName: \"kubernetes.io/projected/7a2cc355-9fd0-4b9b-886c-5d536f714476-kube-api-access-msp87\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.643579 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-config-data\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.643628 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a2cc355-9fd0-4b9b-886c-5d536f714476-run-httpd\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.643960 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.644036 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.644085 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a2cc355-9fd0-4b9b-886c-5d536f714476-log-httpd\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.644152 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.746092 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-scripts\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.746160 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msp87\" (UniqueName: \"kubernetes.io/projected/7a2cc355-9fd0-4b9b-886c-5d536f714476-kube-api-access-msp87\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.746197 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-config-data\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.746280 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a2cc355-9fd0-4b9b-886c-5d536f714476-run-httpd\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.746384 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.746412 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.746445 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a2cc355-9fd0-4b9b-886c-5d536f714476-log-httpd\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.746479 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.747148 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a2cc355-9fd0-4b9b-886c-5d536f714476-run-httpd\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.747323 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a2cc355-9fd0-4b9b-886c-5d536f714476-log-httpd\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.749822 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-scripts\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.750057 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.750071 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.750287 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.751119 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-config-data\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.781767 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msp87\" (UniqueName: \"kubernetes.io/projected/7a2cc355-9fd0-4b9b-886c-5d536f714476-kube-api-access-msp87\") pod \"ceilometer-0\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " pod="openstack/ceilometer-0" Nov 24 17:45:42 crc kubenswrapper[4808]: I1124 17:45:42.856452 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:45:43 crc kubenswrapper[4808]: I1124 17:45:43.298931 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:45:43 crc kubenswrapper[4808]: W1124 17:45:43.306255 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a2cc355_9fd0_4b9b_886c_5d536f714476.slice/crio-d3319523fac505602270f8e75dd8f4950c76a0ec8dfc09a0bc501da6c717cb69 WatchSource:0}: Error finding container d3319523fac505602270f8e75dd8f4950c76a0ec8dfc09a0bc501da6c717cb69: Status 404 returned error can't find the container with id d3319523fac505602270f8e75dd8f4950c76a0ec8dfc09a0bc501da6c717cb69 Nov 24 17:45:44 crc kubenswrapper[4808]: I1124 17:45:44.169005 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a2cc355-9fd0-4b9b-886c-5d536f714476","Type":"ContainerStarted","Data":"d733aea96a428fd56e1963020ccd0f92c5acce9022618ae21096684e59cfc707"} Nov 24 17:45:44 crc kubenswrapper[4808]: I1124 17:45:44.169298 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a2cc355-9fd0-4b9b-886c-5d536f714476","Type":"ContainerStarted","Data":"d3319523fac505602270f8e75dd8f4950c76a0ec8dfc09a0bc501da6c717cb69"} Nov 24 17:45:44 crc kubenswrapper[4808]: I1124 17:45:44.357880 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77896689-86fa-4446-abd2-7c19ffde465c" path="/var/lib/kubelet/pods/77896689-86fa-4446-abd2-7c19ffde465c/volumes" Nov 24 17:45:44 crc kubenswrapper[4808]: I1124 17:45:44.629917 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 17:45:44 crc kubenswrapper[4808]: I1124 17:45:44.664704 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 17:45:44 crc kubenswrapper[4808]: I1124 17:45:44.665305 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 17:45:44 crc kubenswrapper[4808]: I1124 17:45:44.665433 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 17:45:45 crc kubenswrapper[4808]: I1124 17:45:45.179625 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a2cc355-9fd0-4b9b-886c-5d536f714476","Type":"ContainerStarted","Data":"e5e6b376c65c0298f20f7caf4f6e29f5fb9a14786cfbbb7833c9fbbae5342e61"} Nov 24 17:45:45 crc kubenswrapper[4808]: I1124 17:45:45.212564 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 17:45:45 crc kubenswrapper[4808]: I1124 17:45:45.748238 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 17:45:45 crc kubenswrapper[4808]: I1124 17:45:45.748311 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 17:45:46 crc kubenswrapper[4808]: I1124 17:45:46.192802 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a2cc355-9fd0-4b9b-886c-5d536f714476","Type":"ContainerStarted","Data":"19d4b6a3846dd54d43aeee59361510a4e10d7652ab7d9cb373a976c338169171"} Nov 24 17:45:48 crc kubenswrapper[4808]: I1124 17:45:48.210177 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a2cc355-9fd0-4b9b-886c-5d536f714476","Type":"ContainerStarted","Data":"60f2edae3b6c011c46a240c1b1c55bb0ec821101c171bc32f82c46f3fe8c79fe"} Nov 24 17:45:48 crc kubenswrapper[4808]: I1124 17:45:48.210783 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 17:45:48 crc kubenswrapper[4808]: I1124 17:45:48.230843 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8568207810000001 podStartE2EDuration="6.230826639s" podCreationTimestamp="2025-11-24 17:45:42 +0000 UTC" firstStartedPulling="2025-11-24 17:45:43.308788249 +0000 UTC m=+1135.906456051" lastFinishedPulling="2025-11-24 17:45:47.682794117 +0000 UTC m=+1140.280461909" observedRunningTime="2025-11-24 17:45:48.228545354 +0000 UTC m=+1140.826213156" watchObservedRunningTime="2025-11-24 17:45:48.230826639 +0000 UTC m=+1140.828494441" Nov 24 17:45:49 crc kubenswrapper[4808]: I1124 17:45:49.365100 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 17:45:49 crc kubenswrapper[4808]: I1124 17:45:49.369154 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 17:45:49 crc kubenswrapper[4808]: I1124 17:45:49.370845 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 17:45:49 crc kubenswrapper[4808]: I1124 17:45:49.509547 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 17:45:50 crc kubenswrapper[4808]: I1124 17:45:50.233131 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.246210 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.255792 4808 generic.go:334] "Generic (PLEG): container finished" podID="82969a14-94f6-4c08-8d47-b65543a52e80" containerID="2b6e9e6927c329fd49c6eb9a7915a78599d7b187b0e36d07351ade5f98c184fe" exitCode=137 Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.255889 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"82969a14-94f6-4c08-8d47-b65543a52e80","Type":"ContainerDied","Data":"2b6e9e6927c329fd49c6eb9a7915a78599d7b187b0e36d07351ade5f98c184fe"} Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.256397 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"82969a14-94f6-4c08-8d47-b65543a52e80","Type":"ContainerDied","Data":"9b80a6144bf96bea554448f695440df0a3edebb118a73ecdb9ff917470d469cf"} Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.256443 4808 scope.go:117] "RemoveContainer" containerID="2b6e9e6927c329fd49c6eb9a7915a78599d7b187b0e36d07351ade5f98c184fe" Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.291154 4808 scope.go:117] "RemoveContainer" containerID="2b6e9e6927c329fd49c6eb9a7915a78599d7b187b0e36d07351ade5f98c184fe" Nov 24 17:45:52 crc kubenswrapper[4808]: E1124 17:45:52.291812 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b6e9e6927c329fd49c6eb9a7915a78599d7b187b0e36d07351ade5f98c184fe\": container with ID starting with 2b6e9e6927c329fd49c6eb9a7915a78599d7b187b0e36d07351ade5f98c184fe not found: ID does not exist" containerID="2b6e9e6927c329fd49c6eb9a7915a78599d7b187b0e36d07351ade5f98c184fe" Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.291870 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b6e9e6927c329fd49c6eb9a7915a78599d7b187b0e36d07351ade5f98c184fe"} err="failed to get container status \"2b6e9e6927c329fd49c6eb9a7915a78599d7b187b0e36d07351ade5f98c184fe\": rpc error: code = NotFound desc = could not find container \"2b6e9e6927c329fd49c6eb9a7915a78599d7b187b0e36d07351ade5f98c184fe\": container with ID starting with 2b6e9e6927c329fd49c6eb9a7915a78599d7b187b0e36d07351ade5f98c184fe not found: ID does not exist" Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.319102 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46fxj\" (UniqueName: \"kubernetes.io/projected/82969a14-94f6-4c08-8d47-b65543a52e80-kube-api-access-46fxj\") pod \"82969a14-94f6-4c08-8d47-b65543a52e80\" (UID: \"82969a14-94f6-4c08-8d47-b65543a52e80\") " Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.319182 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82969a14-94f6-4c08-8d47-b65543a52e80-config-data\") pod \"82969a14-94f6-4c08-8d47-b65543a52e80\" (UID: \"82969a14-94f6-4c08-8d47-b65543a52e80\") " Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.319412 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82969a14-94f6-4c08-8d47-b65543a52e80-combined-ca-bundle\") pod \"82969a14-94f6-4c08-8d47-b65543a52e80\" (UID: \"82969a14-94f6-4c08-8d47-b65543a52e80\") " Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.341777 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82969a14-94f6-4c08-8d47-b65543a52e80-kube-api-access-46fxj" (OuterVolumeSpecName: "kube-api-access-46fxj") pod "82969a14-94f6-4c08-8d47-b65543a52e80" (UID: "82969a14-94f6-4c08-8d47-b65543a52e80"). InnerVolumeSpecName "kube-api-access-46fxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.378750 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82969a14-94f6-4c08-8d47-b65543a52e80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82969a14-94f6-4c08-8d47-b65543a52e80" (UID: "82969a14-94f6-4c08-8d47-b65543a52e80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.381158 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82969a14-94f6-4c08-8d47-b65543a52e80-config-data" (OuterVolumeSpecName: "config-data") pod "82969a14-94f6-4c08-8d47-b65543a52e80" (UID: "82969a14-94f6-4c08-8d47-b65543a52e80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.421484 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82969a14-94f6-4c08-8d47-b65543a52e80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.421512 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46fxj\" (UniqueName: \"kubernetes.io/projected/82969a14-94f6-4c08-8d47-b65543a52e80-kube-api-access-46fxj\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:52 crc kubenswrapper[4808]: I1124 17:45:52.421537 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82969a14-94f6-4c08-8d47-b65543a52e80-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.276490 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.311164 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.322980 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.333770 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:45:53 crc kubenswrapper[4808]: E1124 17:45:53.334288 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82969a14-94f6-4c08-8d47-b65543a52e80" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.334315 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="82969a14-94f6-4c08-8d47-b65543a52e80" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.334521 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="82969a14-94f6-4c08-8d47-b65543a52e80" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.335504 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.339108 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.339120 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.339164 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.346827 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.437926 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32aec99-9b13-4e2e-af1a-9b55ab86026b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d32aec99-9b13-4e2e-af1a-9b55ab86026b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.438489 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32aec99-9b13-4e2e-af1a-9b55ab86026b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d32aec99-9b13-4e2e-af1a-9b55ab86026b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.439088 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32aec99-9b13-4e2e-af1a-9b55ab86026b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d32aec99-9b13-4e2e-af1a-9b55ab86026b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.439144 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32aec99-9b13-4e2e-af1a-9b55ab86026b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d32aec99-9b13-4e2e-af1a-9b55ab86026b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.439294 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6lj4\" (UniqueName: \"kubernetes.io/projected/d32aec99-9b13-4e2e-af1a-9b55ab86026b-kube-api-access-f6lj4\") pod \"nova-cell1-novncproxy-0\" (UID: \"d32aec99-9b13-4e2e-af1a-9b55ab86026b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.541628 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32aec99-9b13-4e2e-af1a-9b55ab86026b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d32aec99-9b13-4e2e-af1a-9b55ab86026b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.541708 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32aec99-9b13-4e2e-af1a-9b55ab86026b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d32aec99-9b13-4e2e-af1a-9b55ab86026b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.541833 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6lj4\" (UniqueName: \"kubernetes.io/projected/d32aec99-9b13-4e2e-af1a-9b55ab86026b-kube-api-access-f6lj4\") pod \"nova-cell1-novncproxy-0\" (UID: \"d32aec99-9b13-4e2e-af1a-9b55ab86026b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.541883 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32aec99-9b13-4e2e-af1a-9b55ab86026b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d32aec99-9b13-4e2e-af1a-9b55ab86026b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.541917 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32aec99-9b13-4e2e-af1a-9b55ab86026b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d32aec99-9b13-4e2e-af1a-9b55ab86026b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.547830 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32aec99-9b13-4e2e-af1a-9b55ab86026b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d32aec99-9b13-4e2e-af1a-9b55ab86026b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.548860 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32aec99-9b13-4e2e-af1a-9b55ab86026b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d32aec99-9b13-4e2e-af1a-9b55ab86026b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.549421 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32aec99-9b13-4e2e-af1a-9b55ab86026b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d32aec99-9b13-4e2e-af1a-9b55ab86026b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.550477 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32aec99-9b13-4e2e-af1a-9b55ab86026b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d32aec99-9b13-4e2e-af1a-9b55ab86026b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.563115 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6lj4\" (UniqueName: \"kubernetes.io/projected/d32aec99-9b13-4e2e-af1a-9b55ab86026b-kube-api-access-f6lj4\") pod \"nova-cell1-novncproxy-0\" (UID: \"d32aec99-9b13-4e2e-af1a-9b55ab86026b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:53 crc kubenswrapper[4808]: I1124 17:45:53.651365 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:54 crc kubenswrapper[4808]: I1124 17:45:54.133594 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:45:54 crc kubenswrapper[4808]: W1124 17:45:54.137166 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd32aec99_9b13_4e2e_af1a_9b55ab86026b.slice/crio-1b2a45a61396dff08c5c790a5a710437e032c5fb4346a9eca3e778f0129a5997 WatchSource:0}: Error finding container 1b2a45a61396dff08c5c790a5a710437e032c5fb4346a9eca3e778f0129a5997: Status 404 returned error can't find the container with id 1b2a45a61396dff08c5c790a5a710437e032c5fb4346a9eca3e778f0129a5997 Nov 24 17:45:54 crc kubenswrapper[4808]: I1124 17:45:54.287799 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d32aec99-9b13-4e2e-af1a-9b55ab86026b","Type":"ContainerStarted","Data":"1b2a45a61396dff08c5c790a5a710437e032c5fb4346a9eca3e778f0129a5997"} Nov 24 17:45:54 crc kubenswrapper[4808]: I1124 17:45:54.363592 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82969a14-94f6-4c08-8d47-b65543a52e80" path="/var/lib/kubelet/pods/82969a14-94f6-4c08-8d47-b65543a52e80/volumes" Nov 24 17:45:54 crc kubenswrapper[4808]: I1124 17:45:54.669427 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 17:45:54 crc kubenswrapper[4808]: I1124 17:45:54.669895 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 17:45:54 crc kubenswrapper[4808]: I1124 17:45:54.670057 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 17:45:54 crc kubenswrapper[4808]: I1124 17:45:54.675538 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.301530 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d32aec99-9b13-4e2e-af1a-9b55ab86026b","Type":"ContainerStarted","Data":"53816a935ccd0aa58b7c57f69950c35082171b5cef6e1c5fffaa89a11f59935a"} Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.302009 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.306868 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.352652 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.352631783 podStartE2EDuration="2.352631783s" podCreationTimestamp="2025-11-24 17:45:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:45:55.346619162 +0000 UTC m=+1147.944286974" watchObservedRunningTime="2025-11-24 17:45:55.352631783 +0000 UTC m=+1147.950299585" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.561200 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-dzsxl"] Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.565200 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.576641 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-dzsxl"] Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.689981 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.690654 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.691337 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vckkv\" (UniqueName: \"kubernetes.io/projected/378c3582-e95c-4cd0-9ab9-91057bc816ef-kube-api-access-vckkv\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.691470 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-config\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.691668 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.691714 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.793471 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vckkv\" (UniqueName: \"kubernetes.io/projected/378c3582-e95c-4cd0-9ab9-91057bc816ef-kube-api-access-vckkv\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.793544 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-config\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.793621 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.793653 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.793707 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.793735 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.794534 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-config\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.794619 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.794708 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.795462 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.795477 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.820844 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vckkv\" (UniqueName: \"kubernetes.io/projected/378c3582-e95c-4cd0-9ab9-91057bc816ef-kube-api-access-vckkv\") pod \"dnsmasq-dns-89c5cd4d5-dzsxl\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:55 crc kubenswrapper[4808]: I1124 17:45:55.902568 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:56 crc kubenswrapper[4808]: I1124 17:45:56.413159 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-dzsxl"] Nov 24 17:45:57 crc kubenswrapper[4808]: I1124 17:45:57.317900 4808 generic.go:334] "Generic (PLEG): container finished" podID="378c3582-e95c-4cd0-9ab9-91057bc816ef" containerID="fd0e24ce6bb8f63cb56a5b08752d70d0b017ed0e8cb84c4c933a90a00ef9ab9c" exitCode=0 Nov 24 17:45:57 crc kubenswrapper[4808]: I1124 17:45:57.317996 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" event={"ID":"378c3582-e95c-4cd0-9ab9-91057bc816ef","Type":"ContainerDied","Data":"fd0e24ce6bb8f63cb56a5b08752d70d0b017ed0e8cb84c4c933a90a00ef9ab9c"} Nov 24 17:45:57 crc kubenswrapper[4808]: I1124 17:45:57.318308 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" event={"ID":"378c3582-e95c-4cd0-9ab9-91057bc816ef","Type":"ContainerStarted","Data":"a711eab003e74bd0ac484d5a0b2c4abb8c044db0d1bcaf319b6a2c3d8cb9804b"} Nov 24 17:45:57 crc kubenswrapper[4808]: I1124 17:45:57.451075 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:45:57 crc kubenswrapper[4808]: I1124 17:45:57.451328 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="ceilometer-central-agent" containerID="cri-o://d733aea96a428fd56e1963020ccd0f92c5acce9022618ae21096684e59cfc707" gracePeriod=30 Nov 24 17:45:57 crc kubenswrapper[4808]: I1124 17:45:57.451430 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="proxy-httpd" containerID="cri-o://60f2edae3b6c011c46a240c1b1c55bb0ec821101c171bc32f82c46f3fe8c79fe" gracePeriod=30 Nov 24 17:45:57 crc kubenswrapper[4808]: I1124 17:45:57.451464 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="sg-core" containerID="cri-o://19d4b6a3846dd54d43aeee59361510a4e10d7652ab7d9cb373a976c338169171" gracePeriod=30 Nov 24 17:45:57 crc kubenswrapper[4808]: I1124 17:45:57.451501 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="ceilometer-notification-agent" containerID="cri-o://e5e6b376c65c0298f20f7caf4f6e29f5fb9a14786cfbbb7833c9fbbae5342e61" gracePeriod=30 Nov 24 17:45:57 crc kubenswrapper[4808]: I1124 17:45:57.561314 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.203:3000/\": read tcp 10.217.0.2:48392->10.217.0.203:3000: read: connection reset by peer" Nov 24 17:45:57 crc kubenswrapper[4808]: E1124 17:45:57.562989 4808 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a2cc355_9fd0_4b9b_886c_5d536f714476.slice/crio-19d4b6a3846dd54d43aeee59361510a4e10d7652ab7d9cb373a976c338169171.scope\": RecentStats: unable to find data in memory cache]" Nov 24 17:45:57 crc kubenswrapper[4808]: I1124 17:45:57.804746 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:45:58 crc kubenswrapper[4808]: I1124 17:45:58.329652 4808 generic.go:334] "Generic (PLEG): container finished" podID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerID="60f2edae3b6c011c46a240c1b1c55bb0ec821101c171bc32f82c46f3fe8c79fe" exitCode=0 Nov 24 17:45:58 crc kubenswrapper[4808]: I1124 17:45:58.329978 4808 generic.go:334] "Generic (PLEG): container finished" podID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerID="19d4b6a3846dd54d43aeee59361510a4e10d7652ab7d9cb373a976c338169171" exitCode=2 Nov 24 17:45:58 crc kubenswrapper[4808]: I1124 17:45:58.329990 4808 generic.go:334] "Generic (PLEG): container finished" podID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerID="d733aea96a428fd56e1963020ccd0f92c5acce9022618ae21096684e59cfc707" exitCode=0 Nov 24 17:45:58 crc kubenswrapper[4808]: I1124 17:45:58.329724 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a2cc355-9fd0-4b9b-886c-5d536f714476","Type":"ContainerDied","Data":"60f2edae3b6c011c46a240c1b1c55bb0ec821101c171bc32f82c46f3fe8c79fe"} Nov 24 17:45:58 crc kubenswrapper[4808]: I1124 17:45:58.330071 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a2cc355-9fd0-4b9b-886c-5d536f714476","Type":"ContainerDied","Data":"19d4b6a3846dd54d43aeee59361510a4e10d7652ab7d9cb373a976c338169171"} Nov 24 17:45:58 crc kubenswrapper[4808]: I1124 17:45:58.330087 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a2cc355-9fd0-4b9b-886c-5d536f714476","Type":"ContainerDied","Data":"d733aea96a428fd56e1963020ccd0f92c5acce9022618ae21096684e59cfc707"} Nov 24 17:45:58 crc kubenswrapper[4808]: I1124 17:45:58.332381 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" event={"ID":"378c3582-e95c-4cd0-9ab9-91057bc816ef","Type":"ContainerStarted","Data":"fe6ad8a59f01a937082928d3bce3386eb70288ab7a77ab89bc3ad60574cc60ca"} Nov 24 17:45:58 crc kubenswrapper[4808]: I1124 17:45:58.332510 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" containerName="nova-api-log" containerID="cri-o://49b9f183139dfeb7ead25ef7550261b47c0cec64738cdd2156ad1699eb20efec" gracePeriod=30 Nov 24 17:45:58 crc kubenswrapper[4808]: I1124 17:45:58.332779 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" containerName="nova-api-api" containerID="cri-o://2b04932f99541b6d928b2ee93d2391a5c7f583c935c474085c72c7f92e8e729e" gracePeriod=30 Nov 24 17:45:58 crc kubenswrapper[4808]: I1124 17:45:58.333272 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:45:58 crc kubenswrapper[4808]: I1124 17:45:58.358230 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" podStartSLOduration=3.358205861 podStartE2EDuration="3.358205861s" podCreationTimestamp="2025-11-24 17:45:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:45:58.354479785 +0000 UTC m=+1150.952147587" watchObservedRunningTime="2025-11-24 17:45:58.358205861 +0000 UTC m=+1150.955873663" Nov 24 17:45:58 crc kubenswrapper[4808]: I1124 17:45:58.652183 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:45:59 crc kubenswrapper[4808]: I1124 17:45:59.344329 4808 generic.go:334] "Generic (PLEG): container finished" podID="dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" containerID="49b9f183139dfeb7ead25ef7550261b47c0cec64738cdd2156ad1699eb20efec" exitCode=143 Nov 24 17:45:59 crc kubenswrapper[4808]: I1124 17:45:59.344408 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2","Type":"ContainerDied","Data":"49b9f183139dfeb7ead25ef7550261b47c0cec64738cdd2156ad1699eb20efec"} Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.287085 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.370000 4808 generic.go:334] "Generic (PLEG): container finished" podID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerID="e5e6b376c65c0298f20f7caf4f6e29f5fb9a14786cfbbb7833c9fbbae5342e61" exitCode=0 Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.370058 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a2cc355-9fd0-4b9b-886c-5d536f714476","Type":"ContainerDied","Data":"e5e6b376c65c0298f20f7caf4f6e29f5fb9a14786cfbbb7833c9fbbae5342e61"} Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.370084 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a2cc355-9fd0-4b9b-886c-5d536f714476","Type":"ContainerDied","Data":"d3319523fac505602270f8e75dd8f4950c76a0ec8dfc09a0bc501da6c717cb69"} Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.370101 4808 scope.go:117] "RemoveContainer" containerID="60f2edae3b6c011c46a240c1b1c55bb0ec821101c171bc32f82c46f3fe8c79fe" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.370415 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.394572 4808 scope.go:117] "RemoveContainer" containerID="19d4b6a3846dd54d43aeee59361510a4e10d7652ab7d9cb373a976c338169171" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.417969 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-scripts\") pod \"7a2cc355-9fd0-4b9b-886c-5d536f714476\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.418093 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msp87\" (UniqueName: \"kubernetes.io/projected/7a2cc355-9fd0-4b9b-886c-5d536f714476-kube-api-access-msp87\") pod \"7a2cc355-9fd0-4b9b-886c-5d536f714476\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.418146 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a2cc355-9fd0-4b9b-886c-5d536f714476-log-httpd\") pod \"7a2cc355-9fd0-4b9b-886c-5d536f714476\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.418171 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-ceilometer-tls-certs\") pod \"7a2cc355-9fd0-4b9b-886c-5d536f714476\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.418408 4808 scope.go:117] "RemoveContainer" containerID="e5e6b376c65c0298f20f7caf4f6e29f5fb9a14786cfbbb7833c9fbbae5342e61" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.419175 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a2cc355-9fd0-4b9b-886c-5d536f714476-run-httpd\") pod \"7a2cc355-9fd0-4b9b-886c-5d536f714476\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.419215 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-combined-ca-bundle\") pod \"7a2cc355-9fd0-4b9b-886c-5d536f714476\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.419584 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a2cc355-9fd0-4b9b-886c-5d536f714476-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7a2cc355-9fd0-4b9b-886c-5d536f714476" (UID: "7a2cc355-9fd0-4b9b-886c-5d536f714476"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.419775 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a2cc355-9fd0-4b9b-886c-5d536f714476-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7a2cc355-9fd0-4b9b-886c-5d536f714476" (UID: "7a2cc355-9fd0-4b9b-886c-5d536f714476"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.419256 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-sg-core-conf-yaml\") pod \"7a2cc355-9fd0-4b9b-886c-5d536f714476\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.419902 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-config-data\") pod \"7a2cc355-9fd0-4b9b-886c-5d536f714476\" (UID: \"7a2cc355-9fd0-4b9b-886c-5d536f714476\") " Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.422206 4808 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a2cc355-9fd0-4b9b-886c-5d536f714476-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.422806 4808 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a2cc355-9fd0-4b9b-886c-5d536f714476-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.424718 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-scripts" (OuterVolumeSpecName: "scripts") pod "7a2cc355-9fd0-4b9b-886c-5d536f714476" (UID: "7a2cc355-9fd0-4b9b-886c-5d536f714476"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.438412 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a2cc355-9fd0-4b9b-886c-5d536f714476-kube-api-access-msp87" (OuterVolumeSpecName: "kube-api-access-msp87") pod "7a2cc355-9fd0-4b9b-886c-5d536f714476" (UID: "7a2cc355-9fd0-4b9b-886c-5d536f714476"). InnerVolumeSpecName "kube-api-access-msp87". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.450201 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7a2cc355-9fd0-4b9b-886c-5d536f714476" (UID: "7a2cc355-9fd0-4b9b-886c-5d536f714476"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.477741 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "7a2cc355-9fd0-4b9b-886c-5d536f714476" (UID: "7a2cc355-9fd0-4b9b-886c-5d536f714476"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.504726 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a2cc355-9fd0-4b9b-886c-5d536f714476" (UID: "7a2cc355-9fd0-4b9b-886c-5d536f714476"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.523967 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.523994 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msp87\" (UniqueName: \"kubernetes.io/projected/7a2cc355-9fd0-4b9b-886c-5d536f714476-kube-api-access-msp87\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.524006 4808 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.524027 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.524035 4808 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.525489 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-config-data" (OuterVolumeSpecName: "config-data") pod "7a2cc355-9fd0-4b9b-886c-5d536f714476" (UID: "7a2cc355-9fd0-4b9b-886c-5d536f714476"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.626163 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a2cc355-9fd0-4b9b-886c-5d536f714476-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.628812 4808 scope.go:117] "RemoveContainer" containerID="d733aea96a428fd56e1963020ccd0f92c5acce9022618ae21096684e59cfc707" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.646107 4808 scope.go:117] "RemoveContainer" containerID="60f2edae3b6c011c46a240c1b1c55bb0ec821101c171bc32f82c46f3fe8c79fe" Nov 24 17:46:01 crc kubenswrapper[4808]: E1124 17:46:01.646453 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60f2edae3b6c011c46a240c1b1c55bb0ec821101c171bc32f82c46f3fe8c79fe\": container with ID starting with 60f2edae3b6c011c46a240c1b1c55bb0ec821101c171bc32f82c46f3fe8c79fe not found: ID does not exist" containerID="60f2edae3b6c011c46a240c1b1c55bb0ec821101c171bc32f82c46f3fe8c79fe" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.646493 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60f2edae3b6c011c46a240c1b1c55bb0ec821101c171bc32f82c46f3fe8c79fe"} err="failed to get container status \"60f2edae3b6c011c46a240c1b1c55bb0ec821101c171bc32f82c46f3fe8c79fe\": rpc error: code = NotFound desc = could not find container \"60f2edae3b6c011c46a240c1b1c55bb0ec821101c171bc32f82c46f3fe8c79fe\": container with ID starting with 60f2edae3b6c011c46a240c1b1c55bb0ec821101c171bc32f82c46f3fe8c79fe not found: ID does not exist" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.646520 4808 scope.go:117] "RemoveContainer" containerID="19d4b6a3846dd54d43aeee59361510a4e10d7652ab7d9cb373a976c338169171" Nov 24 17:46:01 crc kubenswrapper[4808]: E1124 17:46:01.646783 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19d4b6a3846dd54d43aeee59361510a4e10d7652ab7d9cb373a976c338169171\": container with ID starting with 19d4b6a3846dd54d43aeee59361510a4e10d7652ab7d9cb373a976c338169171 not found: ID does not exist" containerID="19d4b6a3846dd54d43aeee59361510a4e10d7652ab7d9cb373a976c338169171" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.646818 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19d4b6a3846dd54d43aeee59361510a4e10d7652ab7d9cb373a976c338169171"} err="failed to get container status \"19d4b6a3846dd54d43aeee59361510a4e10d7652ab7d9cb373a976c338169171\": rpc error: code = NotFound desc = could not find container \"19d4b6a3846dd54d43aeee59361510a4e10d7652ab7d9cb373a976c338169171\": container with ID starting with 19d4b6a3846dd54d43aeee59361510a4e10d7652ab7d9cb373a976c338169171 not found: ID does not exist" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.646843 4808 scope.go:117] "RemoveContainer" containerID="e5e6b376c65c0298f20f7caf4f6e29f5fb9a14786cfbbb7833c9fbbae5342e61" Nov 24 17:46:01 crc kubenswrapper[4808]: E1124 17:46:01.647069 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5e6b376c65c0298f20f7caf4f6e29f5fb9a14786cfbbb7833c9fbbae5342e61\": container with ID starting with e5e6b376c65c0298f20f7caf4f6e29f5fb9a14786cfbbb7833c9fbbae5342e61 not found: ID does not exist" containerID="e5e6b376c65c0298f20f7caf4f6e29f5fb9a14786cfbbb7833c9fbbae5342e61" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.647093 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5e6b376c65c0298f20f7caf4f6e29f5fb9a14786cfbbb7833c9fbbae5342e61"} err="failed to get container status \"e5e6b376c65c0298f20f7caf4f6e29f5fb9a14786cfbbb7833c9fbbae5342e61\": rpc error: code = NotFound desc = could not find container \"e5e6b376c65c0298f20f7caf4f6e29f5fb9a14786cfbbb7833c9fbbae5342e61\": container with ID starting with e5e6b376c65c0298f20f7caf4f6e29f5fb9a14786cfbbb7833c9fbbae5342e61 not found: ID does not exist" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.647109 4808 scope.go:117] "RemoveContainer" containerID="d733aea96a428fd56e1963020ccd0f92c5acce9022618ae21096684e59cfc707" Nov 24 17:46:01 crc kubenswrapper[4808]: E1124 17:46:01.647341 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d733aea96a428fd56e1963020ccd0f92c5acce9022618ae21096684e59cfc707\": container with ID starting with d733aea96a428fd56e1963020ccd0f92c5acce9022618ae21096684e59cfc707 not found: ID does not exist" containerID="d733aea96a428fd56e1963020ccd0f92c5acce9022618ae21096684e59cfc707" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.647373 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d733aea96a428fd56e1963020ccd0f92c5acce9022618ae21096684e59cfc707"} err="failed to get container status \"d733aea96a428fd56e1963020ccd0f92c5acce9022618ae21096684e59cfc707\": rpc error: code = NotFound desc = could not find container \"d733aea96a428fd56e1963020ccd0f92c5acce9022618ae21096684e59cfc707\": container with ID starting with d733aea96a428fd56e1963020ccd0f92c5acce9022618ae21096684e59cfc707 not found: ID does not exist" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.718150 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.723691 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.745442 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:46:01 crc kubenswrapper[4808]: E1124 17:46:01.746117 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="ceilometer-central-agent" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.746301 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="ceilometer-central-agent" Nov 24 17:46:01 crc kubenswrapper[4808]: E1124 17:46:01.746313 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="sg-core" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.746321 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="sg-core" Nov 24 17:46:01 crc kubenswrapper[4808]: E1124 17:46:01.746355 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="proxy-httpd" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.746363 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="proxy-httpd" Nov 24 17:46:01 crc kubenswrapper[4808]: E1124 17:46:01.746371 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="ceilometer-notification-agent" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.746377 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="ceilometer-notification-agent" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.746745 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="ceilometer-notification-agent" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.746776 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="ceilometer-central-agent" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.749726 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="proxy-httpd" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.749783 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" containerName="sg-core" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.752647 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.756056 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.756321 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.756345 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.760197 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.830003 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-log-httpd\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.830128 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-scripts\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.830157 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whx2b\" (UniqueName: \"kubernetes.io/projected/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-kube-api-access-whx2b\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.830187 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-run-httpd\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.830225 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.830246 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-config-data\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.830266 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.830290 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.931489 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-scripts\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.931545 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whx2b\" (UniqueName: \"kubernetes.io/projected/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-kube-api-access-whx2b\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.931577 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-run-httpd\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.931610 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.931636 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-config-data\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.931657 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.931677 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.931722 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-log-httpd\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.932557 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-run-httpd\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.932707 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-log-httpd\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.937405 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.937469 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-scripts\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.937526 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.938079 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.938746 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-config-data\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:01 crc kubenswrapper[4808]: I1124 17:46:01.950093 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whx2b\" (UniqueName: \"kubernetes.io/projected/a9b7ee10-3d3d-4755-bf68-0d745a8e4c67-kube-api-access-whx2b\") pod \"ceilometer-0\" (UID: \"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67\") " pod="openstack/ceilometer-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.031922 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.078467 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.134293 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-config-data\") pod \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.134674 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-logs\") pod \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.134914 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-combined-ca-bundle\") pod \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.135004 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt6pl\" (UniqueName: \"kubernetes.io/projected/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-kube-api-access-lt6pl\") pod \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\" (UID: \"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2\") " Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.143823 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-logs" (OuterVolumeSpecName: "logs") pod "dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" (UID: "dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.163135 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-kube-api-access-lt6pl" (OuterVolumeSpecName: "kube-api-access-lt6pl") pod "dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" (UID: "dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2"). InnerVolumeSpecName "kube-api-access-lt6pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.186371 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-config-data" (OuterVolumeSpecName: "config-data") pod "dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" (UID: "dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.192210 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" (UID: "dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.247656 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.247685 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt6pl\" (UniqueName: \"kubernetes.io/projected/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-kube-api-access-lt6pl\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.247701 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.247712 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.357468 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a2cc355-9fd0-4b9b-886c-5d536f714476" path="/var/lib/kubelet/pods/7a2cc355-9fd0-4b9b-886c-5d536f714476/volumes" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.388409 4808 generic.go:334] "Generic (PLEG): container finished" podID="dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" containerID="2b04932f99541b6d928b2ee93d2391a5c7f583c935c474085c72c7f92e8e729e" exitCode=0 Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.388451 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2","Type":"ContainerDied","Data":"2b04932f99541b6d928b2ee93d2391a5c7f583c935c474085c72c7f92e8e729e"} Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.388483 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2","Type":"ContainerDied","Data":"27da97426c4ecd0919295d9cd1e0b3ce563520e248b768b3da42222a248b6162"} Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.388501 4808 scope.go:117] "RemoveContainer" containerID="2b04932f99541b6d928b2ee93d2391a5c7f583c935c474085c72c7f92e8e729e" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.388611 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.426800 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.440826 4808 scope.go:117] "RemoveContainer" containerID="49b9f183139dfeb7ead25ef7550261b47c0cec64738cdd2156ad1699eb20efec" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.440962 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.450954 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 17:46:02 crc kubenswrapper[4808]: E1124 17:46:02.451428 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" containerName="nova-api-api" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.451450 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" containerName="nova-api-api" Nov 24 17:46:02 crc kubenswrapper[4808]: E1124 17:46:02.451468 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" containerName="nova-api-log" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.451476 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" containerName="nova-api-log" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.451657 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" containerName="nova-api-api" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.451681 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" containerName="nova-api-log" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.453489 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.461802 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.462177 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.462320 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.465665 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.475814 4808 scope.go:117] "RemoveContainer" containerID="2b04932f99541b6d928b2ee93d2391a5c7f583c935c474085c72c7f92e8e729e" Nov 24 17:46:02 crc kubenswrapper[4808]: E1124 17:46:02.477352 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b04932f99541b6d928b2ee93d2391a5c7f583c935c474085c72c7f92e8e729e\": container with ID starting with 2b04932f99541b6d928b2ee93d2391a5c7f583c935c474085c72c7f92e8e729e not found: ID does not exist" containerID="2b04932f99541b6d928b2ee93d2391a5c7f583c935c474085c72c7f92e8e729e" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.477397 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b04932f99541b6d928b2ee93d2391a5c7f583c935c474085c72c7f92e8e729e"} err="failed to get container status \"2b04932f99541b6d928b2ee93d2391a5c7f583c935c474085c72c7f92e8e729e\": rpc error: code = NotFound desc = could not find container \"2b04932f99541b6d928b2ee93d2391a5c7f583c935c474085c72c7f92e8e729e\": container with ID starting with 2b04932f99541b6d928b2ee93d2391a5c7f583c935c474085c72c7f92e8e729e not found: ID does not exist" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.477426 4808 scope.go:117] "RemoveContainer" containerID="49b9f183139dfeb7ead25ef7550261b47c0cec64738cdd2156ad1699eb20efec" Nov 24 17:46:02 crc kubenswrapper[4808]: E1124 17:46:02.478100 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49b9f183139dfeb7ead25ef7550261b47c0cec64738cdd2156ad1699eb20efec\": container with ID starting with 49b9f183139dfeb7ead25ef7550261b47c0cec64738cdd2156ad1699eb20efec not found: ID does not exist" containerID="49b9f183139dfeb7ead25ef7550261b47c0cec64738cdd2156ad1699eb20efec" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.478152 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49b9f183139dfeb7ead25ef7550261b47c0cec64738cdd2156ad1699eb20efec"} err="failed to get container status \"49b9f183139dfeb7ead25ef7550261b47c0cec64738cdd2156ad1699eb20efec\": rpc error: code = NotFound desc = could not find container \"49b9f183139dfeb7ead25ef7550261b47c0cec64738cdd2156ad1699eb20efec\": container with ID starting with 49b9f183139dfeb7ead25ef7550261b47c0cec64738cdd2156ad1699eb20efec not found: ID does not exist" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.553900 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.553942 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-config-data\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.553974 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-public-tls-certs\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.554075 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-logs\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.554117 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lswvh\" (UniqueName: \"kubernetes.io/projected/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-kube-api-access-lswvh\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.554179 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.595342 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.655598 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.655743 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.655768 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-config-data\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.655796 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-public-tls-certs\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.655843 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-logs\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.655874 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lswvh\" (UniqueName: \"kubernetes.io/projected/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-kube-api-access-lswvh\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.656529 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-logs\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.661006 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.661306 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-public-tls-certs\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.661503 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-config-data\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.661597 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.674748 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lswvh\" (UniqueName: \"kubernetes.io/projected/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-kube-api-access-lswvh\") pod \"nova-api-0\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " pod="openstack/nova-api-0" Nov 24 17:46:02 crc kubenswrapper[4808]: I1124 17:46:02.795769 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:46:03 crc kubenswrapper[4808]: I1124 17:46:03.242364 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:46:03 crc kubenswrapper[4808]: I1124 17:46:03.399876 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67","Type":"ContainerStarted","Data":"9f347bd4e4f5e494bf9ff01cc6ddb9bf927e0ebbd1f48dc406396a444873f7c1"} Nov 24 17:46:03 crc kubenswrapper[4808]: I1124 17:46:03.400205 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67","Type":"ContainerStarted","Data":"f64f462365d909f5b829d14dfbd2f9a852b605275166ef1552f52304ca6fb5d1"} Nov 24 17:46:03 crc kubenswrapper[4808]: I1124 17:46:03.402306 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1","Type":"ContainerStarted","Data":"8e37b1791d48482fed76297d21f123da939af29a01da25153c9710ef3f844c40"} Nov 24 17:46:03 crc kubenswrapper[4808]: I1124 17:46:03.652058 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:46:03 crc kubenswrapper[4808]: I1124 17:46:03.673799 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.364857 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2" path="/var/lib/kubelet/pods/dd6137f4-f5fe-42e3-ae8d-7f25c746e7f2/volumes" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.413530 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67","Type":"ContainerStarted","Data":"150c0a8a1ffb8802b3cbb0ff4f21ce3e01e5a5a947c376987f2bfd14233ed92e"} Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.415260 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1","Type":"ContainerStarted","Data":"82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3"} Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.415307 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1","Type":"ContainerStarted","Data":"f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630"} Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.434277 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.442512 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.442492956 podStartE2EDuration="2.442492956s" podCreationTimestamp="2025-11-24 17:46:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:46:04.439431569 +0000 UTC m=+1157.037099371" watchObservedRunningTime="2025-11-24 17:46:04.442492956 +0000 UTC m=+1157.040160768" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.606709 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-6r66h"] Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.608471 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.610718 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.618066 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.618430 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-6r66h"] Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.703195 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6r66h\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.703541 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-config-data\") pod \"nova-cell1-cell-mapping-6r66h\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.703605 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6cct\" (UniqueName: \"kubernetes.io/projected/3dac3443-b917-41b6-90c9-b23c01524462-kube-api-access-b6cct\") pod \"nova-cell1-cell-mapping-6r66h\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.703641 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-scripts\") pod \"nova-cell1-cell-mapping-6r66h\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.806188 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6r66h\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.806239 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-config-data\") pod \"nova-cell1-cell-mapping-6r66h\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.806289 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6cct\" (UniqueName: \"kubernetes.io/projected/3dac3443-b917-41b6-90c9-b23c01524462-kube-api-access-b6cct\") pod \"nova-cell1-cell-mapping-6r66h\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.806329 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-scripts\") pod \"nova-cell1-cell-mapping-6r66h\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.811337 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-scripts\") pod \"nova-cell1-cell-mapping-6r66h\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.812393 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-config-data\") pod \"nova-cell1-cell-mapping-6r66h\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.818591 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6r66h\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.823198 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6cct\" (UniqueName: \"kubernetes.io/projected/3dac3443-b917-41b6-90c9-b23c01524462-kube-api-access-b6cct\") pod \"nova-cell1-cell-mapping-6r66h\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:04 crc kubenswrapper[4808]: I1124 17:46:04.929783 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:05 crc kubenswrapper[4808]: I1124 17:46:05.427598 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67","Type":"ContainerStarted","Data":"c798e0feb73f143bcb0b21e7dbdb759413832554a8f68a1a2e80cc125947660e"} Nov 24 17:46:05 crc kubenswrapper[4808]: I1124 17:46:05.557926 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-6r66h"] Nov 24 17:46:05 crc kubenswrapper[4808]: W1124 17:46:05.562404 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3dac3443_b917_41b6_90c9_b23c01524462.slice/crio-227396a95ea026a31f466b0a4a3e2db64cb7bc64f7ecb36fbb790e1eede9f715 WatchSource:0}: Error finding container 227396a95ea026a31f466b0a4a3e2db64cb7bc64f7ecb36fbb790e1eede9f715: Status 404 returned error can't find the container with id 227396a95ea026a31f466b0a4a3e2db64cb7bc64f7ecb36fbb790e1eede9f715 Nov 24 17:46:05 crc kubenswrapper[4808]: I1124 17:46:05.904268 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:46:05 crc kubenswrapper[4808]: I1124 17:46:05.980807 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-s2859"] Nov 24 17:46:05 crc kubenswrapper[4808]: I1124 17:46:05.981136 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-s2859" podUID="6b049d97-3af4-4771-8be7-ee521533bce9" containerName="dnsmasq-dns" containerID="cri-o://d6f0b7494cf8448bee489c7f4d06af24466bf0a30c36a8e5227fa2ea327f62e1" gracePeriod=10 Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.434872 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.444185 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6r66h" event={"ID":"3dac3443-b917-41b6-90c9-b23c01524462","Type":"ContainerStarted","Data":"eabf1eccc07ae41bc0daa8a260a65f4b49b7fe8db19eda8e0b5a97ad51312b1d"} Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.444225 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6r66h" event={"ID":"3dac3443-b917-41b6-90c9-b23c01524462","Type":"ContainerStarted","Data":"227396a95ea026a31f466b0a4a3e2db64cb7bc64f7ecb36fbb790e1eede9f715"} Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.450363 4808 generic.go:334] "Generic (PLEG): container finished" podID="6b049d97-3af4-4771-8be7-ee521533bce9" containerID="d6f0b7494cf8448bee489c7f4d06af24466bf0a30c36a8e5227fa2ea327f62e1" exitCode=0 Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.450405 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-s2859" event={"ID":"6b049d97-3af4-4771-8be7-ee521533bce9","Type":"ContainerDied","Data":"d6f0b7494cf8448bee489c7f4d06af24466bf0a30c36a8e5227fa2ea327f62e1"} Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.450429 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-s2859" event={"ID":"6b049d97-3af4-4771-8be7-ee521533bce9","Type":"ContainerDied","Data":"a561329c8ea54407910f27a50aefc674d5f8d5566e192fecc0357879cc024a46"} Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.450444 4808 scope.go:117] "RemoveContainer" containerID="d6f0b7494cf8448bee489c7f4d06af24466bf0a30c36a8e5227fa2ea327f62e1" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.450584 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-s2859" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.490679 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-6r66h" podStartSLOduration=2.4906618099999998 podStartE2EDuration="2.49066181s" podCreationTimestamp="2025-11-24 17:46:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:46:06.483437004 +0000 UTC m=+1159.081104826" watchObservedRunningTime="2025-11-24 17:46:06.49066181 +0000 UTC m=+1159.088329612" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.505372 4808 scope.go:117] "RemoveContainer" containerID="8d33c1dbfc3c22434743ccbb9ae13a0f8b553a3967a885fe22b9b6f72460f6ff" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.522766 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.522828 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.534253 4808 scope.go:117] "RemoveContainer" containerID="d6f0b7494cf8448bee489c7f4d06af24466bf0a30c36a8e5227fa2ea327f62e1" Nov 24 17:46:06 crc kubenswrapper[4808]: E1124 17:46:06.534665 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6f0b7494cf8448bee489c7f4d06af24466bf0a30c36a8e5227fa2ea327f62e1\": container with ID starting with d6f0b7494cf8448bee489c7f4d06af24466bf0a30c36a8e5227fa2ea327f62e1 not found: ID does not exist" containerID="d6f0b7494cf8448bee489c7f4d06af24466bf0a30c36a8e5227fa2ea327f62e1" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.534694 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6f0b7494cf8448bee489c7f4d06af24466bf0a30c36a8e5227fa2ea327f62e1"} err="failed to get container status \"d6f0b7494cf8448bee489c7f4d06af24466bf0a30c36a8e5227fa2ea327f62e1\": rpc error: code = NotFound desc = could not find container \"d6f0b7494cf8448bee489c7f4d06af24466bf0a30c36a8e5227fa2ea327f62e1\": container with ID starting with d6f0b7494cf8448bee489c7f4d06af24466bf0a30c36a8e5227fa2ea327f62e1 not found: ID does not exist" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.534712 4808 scope.go:117] "RemoveContainer" containerID="8d33c1dbfc3c22434743ccbb9ae13a0f8b553a3967a885fe22b9b6f72460f6ff" Nov 24 17:46:06 crc kubenswrapper[4808]: E1124 17:46:06.535068 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d33c1dbfc3c22434743ccbb9ae13a0f8b553a3967a885fe22b9b6f72460f6ff\": container with ID starting with 8d33c1dbfc3c22434743ccbb9ae13a0f8b553a3967a885fe22b9b6f72460f6ff not found: ID does not exist" containerID="8d33c1dbfc3c22434743ccbb9ae13a0f8b553a3967a885fe22b9b6f72460f6ff" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.535095 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d33c1dbfc3c22434743ccbb9ae13a0f8b553a3967a885fe22b9b6f72460f6ff"} err="failed to get container status \"8d33c1dbfc3c22434743ccbb9ae13a0f8b553a3967a885fe22b9b6f72460f6ff\": rpc error: code = NotFound desc = could not find container \"8d33c1dbfc3c22434743ccbb9ae13a0f8b553a3967a885fe22b9b6f72460f6ff\": container with ID starting with 8d33c1dbfc3c22434743ccbb9ae13a0f8b553a3967a885fe22b9b6f72460f6ff not found: ID does not exist" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.537745 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-ovsdbserver-sb\") pod \"6b049d97-3af4-4771-8be7-ee521533bce9\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.537918 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-ovsdbserver-nb\") pod \"6b049d97-3af4-4771-8be7-ee521533bce9\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.537993 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kc5lx\" (UniqueName: \"kubernetes.io/projected/6b049d97-3af4-4771-8be7-ee521533bce9-kube-api-access-kc5lx\") pod \"6b049d97-3af4-4771-8be7-ee521533bce9\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.538095 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-dns-svc\") pod \"6b049d97-3af4-4771-8be7-ee521533bce9\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.538137 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-dns-swift-storage-0\") pod \"6b049d97-3af4-4771-8be7-ee521533bce9\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.538159 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-config\") pod \"6b049d97-3af4-4771-8be7-ee521533bce9\" (UID: \"6b049d97-3af4-4771-8be7-ee521533bce9\") " Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.544161 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b049d97-3af4-4771-8be7-ee521533bce9-kube-api-access-kc5lx" (OuterVolumeSpecName: "kube-api-access-kc5lx") pod "6b049d97-3af4-4771-8be7-ee521533bce9" (UID: "6b049d97-3af4-4771-8be7-ee521533bce9"). InnerVolumeSpecName "kube-api-access-kc5lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.604083 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-config" (OuterVolumeSpecName: "config") pod "6b049d97-3af4-4771-8be7-ee521533bce9" (UID: "6b049d97-3af4-4771-8be7-ee521533bce9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.608624 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6b049d97-3af4-4771-8be7-ee521533bce9" (UID: "6b049d97-3af4-4771-8be7-ee521533bce9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.620313 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6b049d97-3af4-4771-8be7-ee521533bce9" (UID: "6b049d97-3af4-4771-8be7-ee521533bce9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.624275 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6b049d97-3af4-4771-8be7-ee521533bce9" (UID: "6b049d97-3af4-4771-8be7-ee521533bce9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.634463 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6b049d97-3af4-4771-8be7-ee521533bce9" (UID: "6b049d97-3af4-4771-8be7-ee521533bce9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.641657 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kc5lx\" (UniqueName: \"kubernetes.io/projected/6b049d97-3af4-4771-8be7-ee521533bce9-kube-api-access-kc5lx\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.641701 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.641715 4808 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.641724 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.641733 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.641741 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b049d97-3af4-4771-8be7-ee521533bce9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.786357 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-s2859"] Nov 24 17:46:06 crc kubenswrapper[4808]: I1124 17:46:06.801654 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-s2859"] Nov 24 17:46:07 crc kubenswrapper[4808]: I1124 17:46:07.473045 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9b7ee10-3d3d-4755-bf68-0d745a8e4c67","Type":"ContainerStarted","Data":"5fb52c2b03d80872aa383240e4dfc64ced31e07712408e8bbb6cac93309e75d3"} Nov 24 17:46:07 crc kubenswrapper[4808]: I1124 17:46:07.474119 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 17:46:07 crc kubenswrapper[4808]: I1124 17:46:07.498734 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.884978603 podStartE2EDuration="6.49871036s" podCreationTimestamp="2025-11-24 17:46:01 +0000 UTC" firstStartedPulling="2025-11-24 17:46:02.602670281 +0000 UTC m=+1155.200338083" lastFinishedPulling="2025-11-24 17:46:06.216402038 +0000 UTC m=+1158.814069840" observedRunningTime="2025-11-24 17:46:07.495875659 +0000 UTC m=+1160.093543481" watchObservedRunningTime="2025-11-24 17:46:07.49871036 +0000 UTC m=+1160.096378162" Nov 24 17:46:08 crc kubenswrapper[4808]: I1124 17:46:08.360683 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b049d97-3af4-4771-8be7-ee521533bce9" path="/var/lib/kubelet/pods/6b049d97-3af4-4771-8be7-ee521533bce9/volumes" Nov 24 17:46:10 crc kubenswrapper[4808]: I1124 17:46:10.507195 4808 generic.go:334] "Generic (PLEG): container finished" podID="3dac3443-b917-41b6-90c9-b23c01524462" containerID="eabf1eccc07ae41bc0daa8a260a65f4b49b7fe8db19eda8e0b5a97ad51312b1d" exitCode=0 Nov 24 17:46:10 crc kubenswrapper[4808]: I1124 17:46:10.507385 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6r66h" event={"ID":"3dac3443-b917-41b6-90c9-b23c01524462","Type":"ContainerDied","Data":"eabf1eccc07ae41bc0daa8a260a65f4b49b7fe8db19eda8e0b5a97ad51312b1d"} Nov 24 17:46:11 crc kubenswrapper[4808]: I1124 17:46:11.302764 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-757b4f8459-s2859" podUID="6b049d97-3af4-4771-8be7-ee521533bce9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.195:5353: i/o timeout" Nov 24 17:46:11 crc kubenswrapper[4808]: I1124 17:46:11.879670 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:11 crc kubenswrapper[4808]: I1124 17:46:11.943562 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-combined-ca-bundle\") pod \"3dac3443-b917-41b6-90c9-b23c01524462\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " Nov 24 17:46:11 crc kubenswrapper[4808]: I1124 17:46:11.943633 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6cct\" (UniqueName: \"kubernetes.io/projected/3dac3443-b917-41b6-90c9-b23c01524462-kube-api-access-b6cct\") pod \"3dac3443-b917-41b6-90c9-b23c01524462\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " Nov 24 17:46:11 crc kubenswrapper[4808]: I1124 17:46:11.943730 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-scripts\") pod \"3dac3443-b917-41b6-90c9-b23c01524462\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " Nov 24 17:46:11 crc kubenswrapper[4808]: I1124 17:46:11.943940 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-config-data\") pod \"3dac3443-b917-41b6-90c9-b23c01524462\" (UID: \"3dac3443-b917-41b6-90c9-b23c01524462\") " Nov 24 17:46:11 crc kubenswrapper[4808]: I1124 17:46:11.949657 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-scripts" (OuterVolumeSpecName: "scripts") pod "3dac3443-b917-41b6-90c9-b23c01524462" (UID: "3dac3443-b917-41b6-90c9-b23c01524462"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:11 crc kubenswrapper[4808]: I1124 17:46:11.949859 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dac3443-b917-41b6-90c9-b23c01524462-kube-api-access-b6cct" (OuterVolumeSpecName: "kube-api-access-b6cct") pod "3dac3443-b917-41b6-90c9-b23c01524462" (UID: "3dac3443-b917-41b6-90c9-b23c01524462"). InnerVolumeSpecName "kube-api-access-b6cct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:46:11 crc kubenswrapper[4808]: I1124 17:46:11.977977 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3dac3443-b917-41b6-90c9-b23c01524462" (UID: "3dac3443-b917-41b6-90c9-b23c01524462"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:11 crc kubenswrapper[4808]: I1124 17:46:11.983709 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-config-data" (OuterVolumeSpecName: "config-data") pod "3dac3443-b917-41b6-90c9-b23c01524462" (UID: "3dac3443-b917-41b6-90c9-b23c01524462"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:12 crc kubenswrapper[4808]: I1124 17:46:12.046650 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:12 crc kubenswrapper[4808]: I1124 17:46:12.046695 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6cct\" (UniqueName: \"kubernetes.io/projected/3dac3443-b917-41b6-90c9-b23c01524462-kube-api-access-b6cct\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:12 crc kubenswrapper[4808]: I1124 17:46:12.046713 4808 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:12 crc kubenswrapper[4808]: I1124 17:46:12.046725 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dac3443-b917-41b6-90c9-b23c01524462-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:12 crc kubenswrapper[4808]: I1124 17:46:12.525141 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6r66h" Nov 24 17:46:12 crc kubenswrapper[4808]: I1124 17:46:12.525130 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6r66h" event={"ID":"3dac3443-b917-41b6-90c9-b23c01524462","Type":"ContainerDied","Data":"227396a95ea026a31f466b0a4a3e2db64cb7bc64f7ecb36fbb790e1eede9f715"} Nov 24 17:46:12 crc kubenswrapper[4808]: I1124 17:46:12.525192 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="227396a95ea026a31f466b0a4a3e2db64cb7bc64f7ecb36fbb790e1eede9f715" Nov 24 17:46:12 crc kubenswrapper[4808]: I1124 17:46:12.698107 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:46:12 crc kubenswrapper[4808]: I1124 17:46:12.698345 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" containerName="nova-api-log" containerID="cri-o://f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630" gracePeriod=30 Nov 24 17:46:12 crc kubenswrapper[4808]: I1124 17:46:12.698453 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" containerName="nova-api-api" containerID="cri-o://82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3" gracePeriod=30 Nov 24 17:46:12 crc kubenswrapper[4808]: I1124 17:46:12.725086 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:46:12 crc kubenswrapper[4808]: I1124 17:46:12.725678 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2d8bafdb-de51-40bc-b4d2-a4df903ebd38" containerName="nova-scheduler-scheduler" containerID="cri-o://7bff115e81d50a55a8c4cff31b951b1257483600b546b84121a9f4489a0f5ec4" gracePeriod=30 Nov 24 17:46:12 crc kubenswrapper[4808]: I1124 17:46:12.749489 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:46:12 crc kubenswrapper[4808]: I1124 17:46:12.749716 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6359988a-9887-4f69-a3ab-65de38bb758f" containerName="nova-metadata-log" containerID="cri-o://3362edc7aa66c6fa8a37f7b7b785865f30c8b2e749f8de3faf42fd870b6616d7" gracePeriod=30 Nov 24 17:46:12 crc kubenswrapper[4808]: I1124 17:46:12.749834 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6359988a-9887-4f69-a3ab-65de38bb758f" containerName="nova-metadata-metadata" containerID="cri-o://b4db06dd8a5501098c6b11178ae48cf956c33b8e9000b870120d429993d7f2c8" gracePeriod=30 Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.272992 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.371950 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-logs\") pod \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.372126 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-combined-ca-bundle\") pod \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.372162 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-config-data\") pod \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.372253 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lswvh\" (UniqueName: \"kubernetes.io/projected/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-kube-api-access-lswvh\") pod \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.372326 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-internal-tls-certs\") pod \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.372347 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-public-tls-certs\") pod \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\" (UID: \"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1\") " Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.372963 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-logs" (OuterVolumeSpecName: "logs") pod "d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" (UID: "d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.385287 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-kube-api-access-lswvh" (OuterVolumeSpecName: "kube-api-access-lswvh") pod "d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" (UID: "d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1"). InnerVolumeSpecName "kube-api-access-lswvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.404407 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-config-data" (OuterVolumeSpecName: "config-data") pod "d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" (UID: "d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.410948 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" (UID: "d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.431227 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" (UID: "d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.436917 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" (UID: "d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.474846 4808 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.474879 4808 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.474890 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.474899 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.474909 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.474917 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lswvh\" (UniqueName: \"kubernetes.io/projected/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1-kube-api-access-lswvh\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.544762 4808 generic.go:334] "Generic (PLEG): container finished" podID="d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" containerID="82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3" exitCode=0 Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.544803 4808 generic.go:334] "Generic (PLEG): container finished" podID="d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" containerID="f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630" exitCode=143 Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.544854 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1","Type":"ContainerDied","Data":"82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3"} Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.544895 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1","Type":"ContainerDied","Data":"f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630"} Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.544914 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1","Type":"ContainerDied","Data":"8e37b1791d48482fed76297d21f123da939af29a01da25153c9710ef3f844c40"} Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.544932 4808 scope.go:117] "RemoveContainer" containerID="82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.545093 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.552003 4808 generic.go:334] "Generic (PLEG): container finished" podID="6359988a-9887-4f69-a3ab-65de38bb758f" containerID="3362edc7aa66c6fa8a37f7b7b785865f30c8b2e749f8de3faf42fd870b6616d7" exitCode=143 Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.552071 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6359988a-9887-4f69-a3ab-65de38bb758f","Type":"ContainerDied","Data":"3362edc7aa66c6fa8a37f7b7b785865f30c8b2e749f8de3faf42fd870b6616d7"} Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.579825 4808 scope.go:117] "RemoveContainer" containerID="f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.588706 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.601192 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.609998 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 17:46:13 crc kubenswrapper[4808]: E1124 17:46:13.610468 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" containerName="nova-api-api" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.610482 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" containerName="nova-api-api" Nov 24 17:46:13 crc kubenswrapper[4808]: E1124 17:46:13.610502 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b049d97-3af4-4771-8be7-ee521533bce9" containerName="init" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.610509 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b049d97-3af4-4771-8be7-ee521533bce9" containerName="init" Nov 24 17:46:13 crc kubenswrapper[4808]: E1124 17:46:13.610534 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dac3443-b917-41b6-90c9-b23c01524462" containerName="nova-manage" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.610541 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dac3443-b917-41b6-90c9-b23c01524462" containerName="nova-manage" Nov 24 17:46:13 crc kubenswrapper[4808]: E1124 17:46:13.610563 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" containerName="nova-api-log" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.610571 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" containerName="nova-api-log" Nov 24 17:46:13 crc kubenswrapper[4808]: E1124 17:46:13.610584 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b049d97-3af4-4771-8be7-ee521533bce9" containerName="dnsmasq-dns" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.610592 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b049d97-3af4-4771-8be7-ee521533bce9" containerName="dnsmasq-dns" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.610805 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b049d97-3af4-4771-8be7-ee521533bce9" containerName="dnsmasq-dns" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.610821 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dac3443-b917-41b6-90c9-b23c01524462" containerName="nova-manage" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.610833 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" containerName="nova-api-api" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.610847 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" containerName="nova-api-log" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.612620 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.617401 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.626229 4808 scope.go:117] "RemoveContainer" containerID="82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3" Nov 24 17:46:13 crc kubenswrapper[4808]: E1124 17:46:13.627513 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3\": container with ID starting with 82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3 not found: ID does not exist" containerID="82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.627565 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3"} err="failed to get container status \"82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3\": rpc error: code = NotFound desc = could not find container \"82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3\": container with ID starting with 82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3 not found: ID does not exist" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.627591 4808 scope.go:117] "RemoveContainer" containerID="f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630" Nov 24 17:46:13 crc kubenswrapper[4808]: E1124 17:46:13.627898 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630\": container with ID starting with f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630 not found: ID does not exist" containerID="f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.627939 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630"} err="failed to get container status \"f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630\": rpc error: code = NotFound desc = could not find container \"f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630\": container with ID starting with f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630 not found: ID does not exist" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.627958 4808 scope.go:117] "RemoveContainer" containerID="82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.628360 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3"} err="failed to get container status \"82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3\": rpc error: code = NotFound desc = could not find container \"82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3\": container with ID starting with 82a037bc2c8c7309aa728d704144437555beb58fb62a1f8e6ecc675a37599ca3 not found: ID does not exist" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.628407 4808 scope.go:117] "RemoveContainer" containerID="f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.629587 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630"} err="failed to get container status \"f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630\": rpc error: code = NotFound desc = could not find container \"f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630\": container with ID starting with f2894759b0994b1aa943915aa660da23f5c049f937434e60dffa76e599655630 not found: ID does not exist" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.639177 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.643371 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.651509 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.679581 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6861ff-4148-4476-8a97-ef9228b91c8b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.679660 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f6861ff-4148-4476-8a97-ef9228b91c8b-logs\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.679706 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fspw\" (UniqueName: \"kubernetes.io/projected/2f6861ff-4148-4476-8a97-ef9228b91c8b-kube-api-access-6fspw\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.679756 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6861ff-4148-4476-8a97-ef9228b91c8b-config-data\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.679820 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f6861ff-4148-4476-8a97-ef9228b91c8b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.679844 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f6861ff-4148-4476-8a97-ef9228b91c8b-public-tls-certs\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.782067 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f6861ff-4148-4476-8a97-ef9228b91c8b-logs\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.782116 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fspw\" (UniqueName: \"kubernetes.io/projected/2f6861ff-4148-4476-8a97-ef9228b91c8b-kube-api-access-6fspw\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.782160 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6861ff-4148-4476-8a97-ef9228b91c8b-config-data\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.782198 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f6861ff-4148-4476-8a97-ef9228b91c8b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.782215 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f6861ff-4148-4476-8a97-ef9228b91c8b-public-tls-certs\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.782283 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6861ff-4148-4476-8a97-ef9228b91c8b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.784501 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f6861ff-4148-4476-8a97-ef9228b91c8b-logs\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.799253 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6861ff-4148-4476-8a97-ef9228b91c8b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.799601 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f6861ff-4148-4476-8a97-ef9228b91c8b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.800604 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f6861ff-4148-4476-8a97-ef9228b91c8b-public-tls-certs\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.800674 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6861ff-4148-4476-8a97-ef9228b91c8b-config-data\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.816378 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fspw\" (UniqueName: \"kubernetes.io/projected/2f6861ff-4148-4476-8a97-ef9228b91c8b-kube-api-access-6fspw\") pod \"nova-api-0\" (UID: \"2f6861ff-4148-4476-8a97-ef9228b91c8b\") " pod="openstack/nova-api-0" Nov 24 17:46:13 crc kubenswrapper[4808]: I1124 17:46:13.932102 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.052111 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.193602 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-combined-ca-bundle\") pod \"2d8bafdb-de51-40bc-b4d2-a4df903ebd38\" (UID: \"2d8bafdb-de51-40bc-b4d2-a4df903ebd38\") " Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.194133 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwhhw\" (UniqueName: \"kubernetes.io/projected/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-kube-api-access-fwhhw\") pod \"2d8bafdb-de51-40bc-b4d2-a4df903ebd38\" (UID: \"2d8bafdb-de51-40bc-b4d2-a4df903ebd38\") " Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.194203 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-config-data\") pod \"2d8bafdb-de51-40bc-b4d2-a4df903ebd38\" (UID: \"2d8bafdb-de51-40bc-b4d2-a4df903ebd38\") " Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.199237 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-kube-api-access-fwhhw" (OuterVolumeSpecName: "kube-api-access-fwhhw") pod "2d8bafdb-de51-40bc-b4d2-a4df903ebd38" (UID: "2d8bafdb-de51-40bc-b4d2-a4df903ebd38"). InnerVolumeSpecName "kube-api-access-fwhhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.225284 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-config-data" (OuterVolumeSpecName: "config-data") pod "2d8bafdb-de51-40bc-b4d2-a4df903ebd38" (UID: "2d8bafdb-de51-40bc-b4d2-a4df903ebd38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.226143 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d8bafdb-de51-40bc-b4d2-a4df903ebd38" (UID: "2d8bafdb-de51-40bc-b4d2-a4df903ebd38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.296209 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwhhw\" (UniqueName: \"kubernetes.io/projected/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-kube-api-access-fwhhw\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.296240 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.296250 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d8bafdb-de51-40bc-b4d2-a4df903ebd38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.359802 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1" path="/var/lib/kubelet/pods/d05e8bc8-0b5b-4d29-92ba-5092bb3e9bf1/volumes" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.563612 4808 generic.go:334] "Generic (PLEG): container finished" podID="2d8bafdb-de51-40bc-b4d2-a4df903ebd38" containerID="7bff115e81d50a55a8c4cff31b951b1257483600b546b84121a9f4489a0f5ec4" exitCode=0 Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.563680 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2d8bafdb-de51-40bc-b4d2-a4df903ebd38","Type":"ContainerDied","Data":"7bff115e81d50a55a8c4cff31b951b1257483600b546b84121a9f4489a0f5ec4"} Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.563705 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2d8bafdb-de51-40bc-b4d2-a4df903ebd38","Type":"ContainerDied","Data":"046369f04c0758ca2f68c3b3beb962b0ebc4b8107d4b7285e8c40e8ba7225f9f"} Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.563723 4808 scope.go:117] "RemoveContainer" containerID="7bff115e81d50a55a8c4cff31b951b1257483600b546b84121a9f4489a0f5ec4" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.563741 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.586284 4808 scope.go:117] "RemoveContainer" containerID="7bff115e81d50a55a8c4cff31b951b1257483600b546b84121a9f4489a0f5ec4" Nov 24 17:46:14 crc kubenswrapper[4808]: E1124 17:46:14.586794 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bff115e81d50a55a8c4cff31b951b1257483600b546b84121a9f4489a0f5ec4\": container with ID starting with 7bff115e81d50a55a8c4cff31b951b1257483600b546b84121a9f4489a0f5ec4 not found: ID does not exist" containerID="7bff115e81d50a55a8c4cff31b951b1257483600b546b84121a9f4489a0f5ec4" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.586832 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bff115e81d50a55a8c4cff31b951b1257483600b546b84121a9f4489a0f5ec4"} err="failed to get container status \"7bff115e81d50a55a8c4cff31b951b1257483600b546b84121a9f4489a0f5ec4\": rpc error: code = NotFound desc = could not find container \"7bff115e81d50a55a8c4cff31b951b1257483600b546b84121a9f4489a0f5ec4\": container with ID starting with 7bff115e81d50a55a8c4cff31b951b1257483600b546b84121a9f4489a0f5ec4 not found: ID does not exist" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.586870 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.600178 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.606562 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:46:14 crc kubenswrapper[4808]: E1124 17:46:14.606911 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d8bafdb-de51-40bc-b4d2-a4df903ebd38" containerName="nova-scheduler-scheduler" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.606923 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d8bafdb-de51-40bc-b4d2-a4df903ebd38" containerName="nova-scheduler-scheduler" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.607135 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d8bafdb-de51-40bc-b4d2-a4df903ebd38" containerName="nova-scheduler-scheduler" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.608366 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.610511 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.623348 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.703453 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1631f202-c106-4965-874f-7af292288025-config-data\") pod \"nova-scheduler-0\" (UID: \"1631f202-c106-4965-874f-7af292288025\") " pod="openstack/nova-scheduler-0" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.703507 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k64hq\" (UniqueName: \"kubernetes.io/projected/1631f202-c106-4965-874f-7af292288025-kube-api-access-k64hq\") pod \"nova-scheduler-0\" (UID: \"1631f202-c106-4965-874f-7af292288025\") " pod="openstack/nova-scheduler-0" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.703620 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1631f202-c106-4965-874f-7af292288025-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1631f202-c106-4965-874f-7af292288025\") " pod="openstack/nova-scheduler-0" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.805968 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1631f202-c106-4965-874f-7af292288025-config-data\") pod \"nova-scheduler-0\" (UID: \"1631f202-c106-4965-874f-7af292288025\") " pod="openstack/nova-scheduler-0" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.806011 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k64hq\" (UniqueName: \"kubernetes.io/projected/1631f202-c106-4965-874f-7af292288025-kube-api-access-k64hq\") pod \"nova-scheduler-0\" (UID: \"1631f202-c106-4965-874f-7af292288025\") " pod="openstack/nova-scheduler-0" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.806076 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1631f202-c106-4965-874f-7af292288025-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1631f202-c106-4965-874f-7af292288025\") " pod="openstack/nova-scheduler-0" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.809839 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1631f202-c106-4965-874f-7af292288025-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1631f202-c106-4965-874f-7af292288025\") " pod="openstack/nova-scheduler-0" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.810088 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1631f202-c106-4965-874f-7af292288025-config-data\") pod \"nova-scheduler-0\" (UID: \"1631f202-c106-4965-874f-7af292288025\") " pod="openstack/nova-scheduler-0" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.825086 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k64hq\" (UniqueName: \"kubernetes.io/projected/1631f202-c106-4965-874f-7af292288025-kube-api-access-k64hq\") pod \"nova-scheduler-0\" (UID: \"1631f202-c106-4965-874f-7af292288025\") " pod="openstack/nova-scheduler-0" Nov 24 17:46:14 crc kubenswrapper[4808]: I1124 17:46:14.928839 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:46:15 crc kubenswrapper[4808]: I1124 17:46:15.157335 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:46:15 crc kubenswrapper[4808]: I1124 17:46:15.385554 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:46:15 crc kubenswrapper[4808]: W1124 17:46:15.388286 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1631f202_c106_4965_874f_7af292288025.slice/crio-2a2ee7ce997aa925b364f99609753b569bc9eb43187cebce3a52811c676e188a WatchSource:0}: Error finding container 2a2ee7ce997aa925b364f99609753b569bc9eb43187cebce3a52811c676e188a: Status 404 returned error can't find the container with id 2a2ee7ce997aa925b364f99609753b569bc9eb43187cebce3a52811c676e188a Nov 24 17:46:15 crc kubenswrapper[4808]: I1124 17:46:15.576498 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f6861ff-4148-4476-8a97-ef9228b91c8b","Type":"ContainerStarted","Data":"4dc1d4a7f6c213636d7a931f51f745736a86fcbed4aa10cac499de2a579b104d"} Nov 24 17:46:15 crc kubenswrapper[4808]: I1124 17:46:15.576546 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f6861ff-4148-4476-8a97-ef9228b91c8b","Type":"ContainerStarted","Data":"a0429bf968b1e5ee804384d67ef991955f57d8c43d1b239e6a2b7fc67b76e922"} Nov 24 17:46:15 crc kubenswrapper[4808]: I1124 17:46:15.576565 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f6861ff-4148-4476-8a97-ef9228b91c8b","Type":"ContainerStarted","Data":"a063fad83952fb542e943701e9659ce11a286ba033bd40b468a7fa15e9b18d5c"} Nov 24 17:46:15 crc kubenswrapper[4808]: I1124 17:46:15.580067 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1631f202-c106-4965-874f-7af292288025","Type":"ContainerStarted","Data":"138d23454b747bf0cf68e98208f6ff1d82c249be2f352ca9c6b5f0e2e6a249d1"} Nov 24 17:46:15 crc kubenswrapper[4808]: I1124 17:46:15.580124 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1631f202-c106-4965-874f-7af292288025","Type":"ContainerStarted","Data":"2a2ee7ce997aa925b364f99609753b569bc9eb43187cebce3a52811c676e188a"} Nov 24 17:46:15 crc kubenswrapper[4808]: I1124 17:46:15.596264 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.596245092 podStartE2EDuration="2.596245092s" podCreationTimestamp="2025-11-24 17:46:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:46:15.593348669 +0000 UTC m=+1168.191016491" watchObservedRunningTime="2025-11-24 17:46:15.596245092 +0000 UTC m=+1168.193912894" Nov 24 17:46:15 crc kubenswrapper[4808]: I1124 17:46:15.618847 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.6188263470000002 podStartE2EDuration="1.618826347s" podCreationTimestamp="2025-11-24 17:46:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:46:15.606362431 +0000 UTC m=+1168.204030253" watchObservedRunningTime="2025-11-24 17:46:15.618826347 +0000 UTC m=+1168.216494149" Nov 24 17:46:15 crc kubenswrapper[4808]: I1124 17:46:15.881472 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6359988a-9887-4f69-a3ab-65de38bb758f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": read tcp 10.217.0.2:54884->10.217.0.198:8775: read: connection reset by peer" Nov 24 17:46:15 crc kubenswrapper[4808]: I1124 17:46:15.881834 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6359988a-9887-4f69-a3ab-65de38bb758f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": read tcp 10.217.0.2:54898->10.217.0.198:8775: read: connection reset by peer" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.340175 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.367051 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d8bafdb-de51-40bc-b4d2-a4df903ebd38" path="/var/lib/kubelet/pods/2d8bafdb-de51-40bc-b4d2-a4df903ebd38/volumes" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.436106 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6359988a-9887-4f69-a3ab-65de38bb758f-logs\") pod \"6359988a-9887-4f69-a3ab-65de38bb758f\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.436518 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-combined-ca-bundle\") pod \"6359988a-9887-4f69-a3ab-65de38bb758f\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.436653 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-nova-metadata-tls-certs\") pod \"6359988a-9887-4f69-a3ab-65de38bb758f\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.436683 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsdtm\" (UniqueName: \"kubernetes.io/projected/6359988a-9887-4f69-a3ab-65de38bb758f-kube-api-access-xsdtm\") pod \"6359988a-9887-4f69-a3ab-65de38bb758f\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.436712 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-config-data\") pod \"6359988a-9887-4f69-a3ab-65de38bb758f\" (UID: \"6359988a-9887-4f69-a3ab-65de38bb758f\") " Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.436946 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6359988a-9887-4f69-a3ab-65de38bb758f-logs" (OuterVolumeSpecName: "logs") pod "6359988a-9887-4f69-a3ab-65de38bb758f" (UID: "6359988a-9887-4f69-a3ab-65de38bb758f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.438045 4808 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6359988a-9887-4f69-a3ab-65de38bb758f-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.445300 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6359988a-9887-4f69-a3ab-65de38bb758f-kube-api-access-xsdtm" (OuterVolumeSpecName: "kube-api-access-xsdtm") pod "6359988a-9887-4f69-a3ab-65de38bb758f" (UID: "6359988a-9887-4f69-a3ab-65de38bb758f"). InnerVolumeSpecName "kube-api-access-xsdtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.478471 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-config-data" (OuterVolumeSpecName: "config-data") pod "6359988a-9887-4f69-a3ab-65de38bb758f" (UID: "6359988a-9887-4f69-a3ab-65de38bb758f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.491285 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6359988a-9887-4f69-a3ab-65de38bb758f" (UID: "6359988a-9887-4f69-a3ab-65de38bb758f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.527940 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6359988a-9887-4f69-a3ab-65de38bb758f" (UID: "6359988a-9887-4f69-a3ab-65de38bb758f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.539794 4808 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.539833 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsdtm\" (UniqueName: \"kubernetes.io/projected/6359988a-9887-4f69-a3ab-65de38bb758f-kube-api-access-xsdtm\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.539843 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.539853 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6359988a-9887-4f69-a3ab-65de38bb758f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.590152 4808 generic.go:334] "Generic (PLEG): container finished" podID="6359988a-9887-4f69-a3ab-65de38bb758f" containerID="b4db06dd8a5501098c6b11178ae48cf956c33b8e9000b870120d429993d7f2c8" exitCode=0 Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.590983 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.594422 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6359988a-9887-4f69-a3ab-65de38bb758f","Type":"ContainerDied","Data":"b4db06dd8a5501098c6b11178ae48cf956c33b8e9000b870120d429993d7f2c8"} Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.594508 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6359988a-9887-4f69-a3ab-65de38bb758f","Type":"ContainerDied","Data":"8f48392722cb1b921279be478b98fe5092bfb19dd190473ae40e05fd58d27628"} Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.594539 4808 scope.go:117] "RemoveContainer" containerID="b4db06dd8a5501098c6b11178ae48cf956c33b8e9000b870120d429993d7f2c8" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.626429 4808 scope.go:117] "RemoveContainer" containerID="3362edc7aa66c6fa8a37f7b7b785865f30c8b2e749f8de3faf42fd870b6616d7" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.645849 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.662879 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.669117 4808 scope.go:117] "RemoveContainer" containerID="b4db06dd8a5501098c6b11178ae48cf956c33b8e9000b870120d429993d7f2c8" Nov 24 17:46:16 crc kubenswrapper[4808]: E1124 17:46:16.669757 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4db06dd8a5501098c6b11178ae48cf956c33b8e9000b870120d429993d7f2c8\": container with ID starting with b4db06dd8a5501098c6b11178ae48cf956c33b8e9000b870120d429993d7f2c8 not found: ID does not exist" containerID="b4db06dd8a5501098c6b11178ae48cf956c33b8e9000b870120d429993d7f2c8" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.669798 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4db06dd8a5501098c6b11178ae48cf956c33b8e9000b870120d429993d7f2c8"} err="failed to get container status \"b4db06dd8a5501098c6b11178ae48cf956c33b8e9000b870120d429993d7f2c8\": rpc error: code = NotFound desc = could not find container \"b4db06dd8a5501098c6b11178ae48cf956c33b8e9000b870120d429993d7f2c8\": container with ID starting with b4db06dd8a5501098c6b11178ae48cf956c33b8e9000b870120d429993d7f2c8 not found: ID does not exist" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.669825 4808 scope.go:117] "RemoveContainer" containerID="3362edc7aa66c6fa8a37f7b7b785865f30c8b2e749f8de3faf42fd870b6616d7" Nov 24 17:46:16 crc kubenswrapper[4808]: E1124 17:46:16.670869 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3362edc7aa66c6fa8a37f7b7b785865f30c8b2e749f8de3faf42fd870b6616d7\": container with ID starting with 3362edc7aa66c6fa8a37f7b7b785865f30c8b2e749f8de3faf42fd870b6616d7 not found: ID does not exist" containerID="3362edc7aa66c6fa8a37f7b7b785865f30c8b2e749f8de3faf42fd870b6616d7" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.670897 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3362edc7aa66c6fa8a37f7b7b785865f30c8b2e749f8de3faf42fd870b6616d7"} err="failed to get container status \"3362edc7aa66c6fa8a37f7b7b785865f30c8b2e749f8de3faf42fd870b6616d7\": rpc error: code = NotFound desc = could not find container \"3362edc7aa66c6fa8a37f7b7b785865f30c8b2e749f8de3faf42fd870b6616d7\": container with ID starting with 3362edc7aa66c6fa8a37f7b7b785865f30c8b2e749f8de3faf42fd870b6616d7 not found: ID does not exist" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.686706 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:46:16 crc kubenswrapper[4808]: E1124 17:46:16.687264 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6359988a-9887-4f69-a3ab-65de38bb758f" containerName="nova-metadata-metadata" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.687282 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6359988a-9887-4f69-a3ab-65de38bb758f" containerName="nova-metadata-metadata" Nov 24 17:46:16 crc kubenswrapper[4808]: E1124 17:46:16.687315 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6359988a-9887-4f69-a3ab-65de38bb758f" containerName="nova-metadata-log" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.687326 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6359988a-9887-4f69-a3ab-65de38bb758f" containerName="nova-metadata-log" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.687590 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="6359988a-9887-4f69-a3ab-65de38bb758f" containerName="nova-metadata-metadata" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.687620 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="6359988a-9887-4f69-a3ab-65de38bb758f" containerName="nova-metadata-log" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.688931 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.694922 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.695500 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.698713 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.746541 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e12cb986-dbb8-42c6-a6f4-7d49d49107d6-config-data\") pod \"nova-metadata-0\" (UID: \"e12cb986-dbb8-42c6-a6f4-7d49d49107d6\") " pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.746618 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e12cb986-dbb8-42c6-a6f4-7d49d49107d6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e12cb986-dbb8-42c6-a6f4-7d49d49107d6\") " pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.746848 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-745jq\" (UniqueName: \"kubernetes.io/projected/e12cb986-dbb8-42c6-a6f4-7d49d49107d6-kube-api-access-745jq\") pod \"nova-metadata-0\" (UID: \"e12cb986-dbb8-42c6-a6f4-7d49d49107d6\") " pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.747185 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e12cb986-dbb8-42c6-a6f4-7d49d49107d6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e12cb986-dbb8-42c6-a6f4-7d49d49107d6\") " pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.747396 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e12cb986-dbb8-42c6-a6f4-7d49d49107d6-logs\") pod \"nova-metadata-0\" (UID: \"e12cb986-dbb8-42c6-a6f4-7d49d49107d6\") " pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.848608 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e12cb986-dbb8-42c6-a6f4-7d49d49107d6-config-data\") pod \"nova-metadata-0\" (UID: \"e12cb986-dbb8-42c6-a6f4-7d49d49107d6\") " pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.848666 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e12cb986-dbb8-42c6-a6f4-7d49d49107d6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e12cb986-dbb8-42c6-a6f4-7d49d49107d6\") " pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.848710 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-745jq\" (UniqueName: \"kubernetes.io/projected/e12cb986-dbb8-42c6-a6f4-7d49d49107d6-kube-api-access-745jq\") pod \"nova-metadata-0\" (UID: \"e12cb986-dbb8-42c6-a6f4-7d49d49107d6\") " pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.848831 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e12cb986-dbb8-42c6-a6f4-7d49d49107d6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e12cb986-dbb8-42c6-a6f4-7d49d49107d6\") " pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.848903 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e12cb986-dbb8-42c6-a6f4-7d49d49107d6-logs\") pod \"nova-metadata-0\" (UID: \"e12cb986-dbb8-42c6-a6f4-7d49d49107d6\") " pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.849390 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e12cb986-dbb8-42c6-a6f4-7d49d49107d6-logs\") pod \"nova-metadata-0\" (UID: \"e12cb986-dbb8-42c6-a6f4-7d49d49107d6\") " pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.852996 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e12cb986-dbb8-42c6-a6f4-7d49d49107d6-config-data\") pod \"nova-metadata-0\" (UID: \"e12cb986-dbb8-42c6-a6f4-7d49d49107d6\") " pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.853316 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e12cb986-dbb8-42c6-a6f4-7d49d49107d6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e12cb986-dbb8-42c6-a6f4-7d49d49107d6\") " pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.853815 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e12cb986-dbb8-42c6-a6f4-7d49d49107d6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e12cb986-dbb8-42c6-a6f4-7d49d49107d6\") " pod="openstack/nova-metadata-0" Nov 24 17:46:16 crc kubenswrapper[4808]: I1124 17:46:16.869034 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-745jq\" (UniqueName: \"kubernetes.io/projected/e12cb986-dbb8-42c6-a6f4-7d49d49107d6-kube-api-access-745jq\") pod \"nova-metadata-0\" (UID: \"e12cb986-dbb8-42c6-a6f4-7d49d49107d6\") " pod="openstack/nova-metadata-0" Nov 24 17:46:17 crc kubenswrapper[4808]: I1124 17:46:17.018926 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:46:17 crc kubenswrapper[4808]: I1124 17:46:17.479483 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:46:17 crc kubenswrapper[4808]: W1124 17:46:17.483846 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode12cb986_dbb8_42c6_a6f4_7d49d49107d6.slice/crio-3a88389605ef356553079bad730c071310fd669750dd405a7748b49d404537f5 WatchSource:0}: Error finding container 3a88389605ef356553079bad730c071310fd669750dd405a7748b49d404537f5: Status 404 returned error can't find the container with id 3a88389605ef356553079bad730c071310fd669750dd405a7748b49d404537f5 Nov 24 17:46:17 crc kubenswrapper[4808]: I1124 17:46:17.601914 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e12cb986-dbb8-42c6-a6f4-7d49d49107d6","Type":"ContainerStarted","Data":"3a88389605ef356553079bad730c071310fd669750dd405a7748b49d404537f5"} Nov 24 17:46:18 crc kubenswrapper[4808]: I1124 17:46:18.367661 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6359988a-9887-4f69-a3ab-65de38bb758f" path="/var/lib/kubelet/pods/6359988a-9887-4f69-a3ab-65de38bb758f/volumes" Nov 24 17:46:18 crc kubenswrapper[4808]: I1124 17:46:18.613924 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e12cb986-dbb8-42c6-a6f4-7d49d49107d6","Type":"ContainerStarted","Data":"0b2ccd1a61743c8f6d3ff40cf090fb004bf0be042243bf763d50e02eeb50fcaf"} Nov 24 17:46:18 crc kubenswrapper[4808]: I1124 17:46:18.614108 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e12cb986-dbb8-42c6-a6f4-7d49d49107d6","Type":"ContainerStarted","Data":"0cf3cbcaf38923a108ba8e2e2657958608adfa8a0c9f7e54251d22b910af897f"} Nov 24 17:46:18 crc kubenswrapper[4808]: I1124 17:46:18.648472 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.648450781 podStartE2EDuration="2.648450781s" podCreationTimestamp="2025-11-24 17:46:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:46:18.639423133 +0000 UTC m=+1171.237090945" watchObservedRunningTime="2025-11-24 17:46:18.648450781 +0000 UTC m=+1171.246118583" Nov 24 17:46:19 crc kubenswrapper[4808]: I1124 17:46:19.929378 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 17:46:22 crc kubenswrapper[4808]: I1124 17:46:22.019128 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 17:46:22 crc kubenswrapper[4808]: I1124 17:46:22.019451 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 17:46:23 crc kubenswrapper[4808]: I1124 17:46:23.933696 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 17:46:23 crc kubenswrapper[4808]: I1124 17:46:23.935361 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 17:46:24 crc kubenswrapper[4808]: I1124 17:46:24.929557 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 17:46:24 crc kubenswrapper[4808]: I1124 17:46:24.946333 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f6861ff-4148-4476-8a97-ef9228b91c8b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 17:46:24 crc kubenswrapper[4808]: I1124 17:46:24.946390 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f6861ff-4148-4476-8a97-ef9228b91c8b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 17:46:24 crc kubenswrapper[4808]: I1124 17:46:24.965159 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 17:46:25 crc kubenswrapper[4808]: I1124 17:46:25.713437 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 17:46:27 crc kubenswrapper[4808]: I1124 17:46:27.019244 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 17:46:27 crc kubenswrapper[4808]: I1124 17:46:27.019650 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 17:46:28 crc kubenswrapper[4808]: I1124 17:46:28.035211 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e12cb986-dbb8-42c6-a6f4-7d49d49107d6" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 17:46:28 crc kubenswrapper[4808]: I1124 17:46:28.035232 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e12cb986-dbb8-42c6-a6f4-7d49d49107d6" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 17:46:32 crc kubenswrapper[4808]: I1124 17:46:32.086256 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 17:46:33 crc kubenswrapper[4808]: I1124 17:46:33.937831 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 17:46:33 crc kubenswrapper[4808]: I1124 17:46:33.938451 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 17:46:33 crc kubenswrapper[4808]: I1124 17:46:33.939298 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 17:46:33 crc kubenswrapper[4808]: I1124 17:46:33.944841 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 17:46:34 crc kubenswrapper[4808]: I1124 17:46:34.769060 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 17:46:34 crc kubenswrapper[4808]: I1124 17:46:34.776482 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 17:46:36 crc kubenswrapper[4808]: I1124 17:46:36.523255 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:46:36 crc kubenswrapper[4808]: I1124 17:46:36.523566 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:46:37 crc kubenswrapper[4808]: I1124 17:46:37.024511 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 17:46:37 crc kubenswrapper[4808]: I1124 17:46:37.024886 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 17:46:37 crc kubenswrapper[4808]: I1124 17:46:37.028382 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 17:46:37 crc kubenswrapper[4808]: I1124 17:46:37.797225 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 17:46:45 crc kubenswrapper[4808]: I1124 17:46:45.805324 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:46:47 crc kubenswrapper[4808]: I1124 17:46:47.099819 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:46:50 crc kubenswrapper[4808]: I1124 17:46:50.465864 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="e0b9711c-4309-40aa-8902-e2ad54e47000" containerName="rabbitmq" containerID="cri-o://34753fc59dc288446ec80afdcfddee939d496ccc71ebc89ccf085eeac368798c" gracePeriod=604796 Nov 24 17:46:50 crc kubenswrapper[4808]: I1124 17:46:50.658699 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="e0b9711c-4309-40aa-8902-e2ad54e47000" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Nov 24 17:46:51 crc kubenswrapper[4808]: I1124 17:46:51.562236 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="c03d629c-1abf-440f-b3c8-d70e93e36eb1" containerName="rabbitmq" containerID="cri-o://fc9eaa638279c561b3f3ba4fdf1622a85a985a4623528792cf646cc044e0b0c3" gracePeriod=604796 Nov 24 17:46:56 crc kubenswrapper[4808]: I1124 17:46:56.958762 4808 generic.go:334] "Generic (PLEG): container finished" podID="e0b9711c-4309-40aa-8902-e2ad54e47000" containerID="34753fc59dc288446ec80afdcfddee939d496ccc71ebc89ccf085eeac368798c" exitCode=0 Nov 24 17:46:56 crc kubenswrapper[4808]: I1124 17:46:56.958825 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e0b9711c-4309-40aa-8902-e2ad54e47000","Type":"ContainerDied","Data":"34753fc59dc288446ec80afdcfddee939d496ccc71ebc89ccf085eeac368798c"} Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.384979 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.503353 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-plugins-conf\") pod \"e0b9711c-4309-40aa-8902-e2ad54e47000\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.503413 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e0b9711c-4309-40aa-8902-e2ad54e47000-erlang-cookie-secret\") pod \"e0b9711c-4309-40aa-8902-e2ad54e47000\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.503478 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-config-data\") pod \"e0b9711c-4309-40aa-8902-e2ad54e47000\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.503499 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-confd\") pod \"e0b9711c-4309-40aa-8902-e2ad54e47000\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.503526 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-plugins\") pod \"e0b9711c-4309-40aa-8902-e2ad54e47000\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.503558 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-erlang-cookie\") pod \"e0b9711c-4309-40aa-8902-e2ad54e47000\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.503578 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e0b9711c-4309-40aa-8902-e2ad54e47000-pod-info\") pod \"e0b9711c-4309-40aa-8902-e2ad54e47000\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.503598 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-server-conf\") pod \"e0b9711c-4309-40aa-8902-e2ad54e47000\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.503626 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"e0b9711c-4309-40aa-8902-e2ad54e47000\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.503767 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-tls\") pod \"e0b9711c-4309-40aa-8902-e2ad54e47000\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.503794 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vppw8\" (UniqueName: \"kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-kube-api-access-vppw8\") pod \"e0b9711c-4309-40aa-8902-e2ad54e47000\" (UID: \"e0b9711c-4309-40aa-8902-e2ad54e47000\") " Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.505206 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "e0b9711c-4309-40aa-8902-e2ad54e47000" (UID: "e0b9711c-4309-40aa-8902-e2ad54e47000"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.505534 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "e0b9711c-4309-40aa-8902-e2ad54e47000" (UID: "e0b9711c-4309-40aa-8902-e2ad54e47000"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.505637 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "e0b9711c-4309-40aa-8902-e2ad54e47000" (UID: "e0b9711c-4309-40aa-8902-e2ad54e47000"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.510673 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-kube-api-access-vppw8" (OuterVolumeSpecName: "kube-api-access-vppw8") pod "e0b9711c-4309-40aa-8902-e2ad54e47000" (UID: "e0b9711c-4309-40aa-8902-e2ad54e47000"). InnerVolumeSpecName "kube-api-access-vppw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.545291 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0b9711c-4309-40aa-8902-e2ad54e47000-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "e0b9711c-4309-40aa-8902-e2ad54e47000" (UID: "e0b9711c-4309-40aa-8902-e2ad54e47000"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.545819 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "e0b9711c-4309-40aa-8902-e2ad54e47000" (UID: "e0b9711c-4309-40aa-8902-e2ad54e47000"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.545852 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/e0b9711c-4309-40aa-8902-e2ad54e47000-pod-info" (OuterVolumeSpecName: "pod-info") pod "e0b9711c-4309-40aa-8902-e2ad54e47000" (UID: "e0b9711c-4309-40aa-8902-e2ad54e47000"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.547442 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "e0b9711c-4309-40aa-8902-e2ad54e47000" (UID: "e0b9711c-4309-40aa-8902-e2ad54e47000"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.551631 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-config-data" (OuterVolumeSpecName: "config-data") pod "e0b9711c-4309-40aa-8902-e2ad54e47000" (UID: "e0b9711c-4309-40aa-8902-e2ad54e47000"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.573204 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-server-conf" (OuterVolumeSpecName: "server-conf") pod "e0b9711c-4309-40aa-8902-e2ad54e47000" (UID: "e0b9711c-4309-40aa-8902-e2ad54e47000"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.612785 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.612847 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vppw8\" (UniqueName: \"kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-kube-api-access-vppw8\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.612865 4808 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.612876 4808 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e0b9711c-4309-40aa-8902-e2ad54e47000-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.612886 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.612897 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.612928 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.612938 4808 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e0b9711c-4309-40aa-8902-e2ad54e47000-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.612960 4808 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e0b9711c-4309-40aa-8902-e2ad54e47000-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.612984 4808 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.628448 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "e0b9711c-4309-40aa-8902-e2ad54e47000" (UID: "e0b9711c-4309-40aa-8902-e2ad54e47000"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.649898 4808 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.714504 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e0b9711c-4309-40aa-8902-e2ad54e47000-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.714537 4808 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.756308 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-txw9q"] Nov 24 17:46:57 crc kubenswrapper[4808]: E1124 17:46:57.756670 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0b9711c-4309-40aa-8902-e2ad54e47000" containerName="rabbitmq" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.756686 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0b9711c-4309-40aa-8902-e2ad54e47000" containerName="rabbitmq" Nov 24 17:46:57 crc kubenswrapper[4808]: E1124 17:46:57.756712 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0b9711c-4309-40aa-8902-e2ad54e47000" containerName="setup-container" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.756719 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0b9711c-4309-40aa-8902-e2ad54e47000" containerName="setup-container" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.756890 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0b9711c-4309-40aa-8902-e2ad54e47000" containerName="rabbitmq" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.757893 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.763295 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.775817 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-txw9q"] Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.816779 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-config\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.816920 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.816950 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.816992 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8thjb\" (UniqueName: \"kubernetes.io/projected/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-kube-api-access-8thjb\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.817041 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.817068 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.817277 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.919401 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.919536 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-config\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.919646 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.919667 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.919703 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8thjb\" (UniqueName: \"kubernetes.io/projected/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-kube-api-access-8thjb\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.919731 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.919763 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.920751 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.921320 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.921590 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.921973 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.922574 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-config\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.922629 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.940065 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8thjb\" (UniqueName: \"kubernetes.io/projected/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-kube-api-access-8thjb\") pod \"dnsmasq-dns-79bd4cc8c9-txw9q\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.970798 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e0b9711c-4309-40aa-8902-e2ad54e47000","Type":"ContainerDied","Data":"7eb981fb62552051609b91fd0714ad6129273494c616ef4089769ad6c9856132"} Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.970868 4808 scope.go:117] "RemoveContainer" containerID="34753fc59dc288446ec80afdcfddee939d496ccc71ebc89ccf085eeac368798c" Nov 24 17:46:57 crc kubenswrapper[4808]: I1124 17:46:57.970871 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.007717 4808 scope.go:117] "RemoveContainer" containerID="417e38f4fcc06e5f31f575f1e377136dd887c214621183abc1e021be79950f22" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.014076 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.026754 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.041622 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.043282 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.052290 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.052545 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.052781 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.052786 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.052934 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.053003 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-rkkz4" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.053352 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.074215 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.076034 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.122809 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bddeab55-cf79-4073-b5c4-420508603e10-config-data\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.122864 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.122937 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bddeab55-cf79-4073-b5c4-420508603e10-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.122966 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bddeab55-cf79-4073-b5c4-420508603e10-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.122994 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bddeab55-cf79-4073-b5c4-420508603e10-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.123030 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bddeab55-cf79-4073-b5c4-420508603e10-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.123051 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h65f\" (UniqueName: \"kubernetes.io/projected/bddeab55-cf79-4073-b5c4-420508603e10-kube-api-access-9h65f\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.123110 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bddeab55-cf79-4073-b5c4-420508603e10-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.123160 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bddeab55-cf79-4073-b5c4-420508603e10-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.123197 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bddeab55-cf79-4073-b5c4-420508603e10-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.123290 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bddeab55-cf79-4073-b5c4-420508603e10-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.228212 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bddeab55-cf79-4073-b5c4-420508603e10-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.225588 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bddeab55-cf79-4073-b5c4-420508603e10-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.230093 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bddeab55-cf79-4073-b5c4-420508603e10-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.230150 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h65f\" (UniqueName: \"kubernetes.io/projected/bddeab55-cf79-4073-b5c4-420508603e10-kube-api-access-9h65f\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.230200 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bddeab55-cf79-4073-b5c4-420508603e10-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.230332 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bddeab55-cf79-4073-b5c4-420508603e10-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.233601 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bddeab55-cf79-4073-b5c4-420508603e10-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.233730 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bddeab55-cf79-4073-b5c4-420508603e10-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.233798 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bddeab55-cf79-4073-b5c4-420508603e10-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.233860 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bddeab55-cf79-4073-b5c4-420508603e10-config-data\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.233895 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.233975 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bddeab55-cf79-4073-b5c4-420508603e10-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.234042 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bddeab55-cf79-4073-b5c4-420508603e10-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.235831 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bddeab55-cf79-4073-b5c4-420508603e10-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.235983 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bddeab55-cf79-4073-b5c4-420508603e10-config-data\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.237061 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bddeab55-cf79-4073-b5c4-420508603e10-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.237190 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.240517 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bddeab55-cf79-4073-b5c4-420508603e10-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.242804 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bddeab55-cf79-4073-b5c4-420508603e10-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.245936 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bddeab55-cf79-4073-b5c4-420508603e10-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.248539 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h65f\" (UniqueName: \"kubernetes.io/projected/bddeab55-cf79-4073-b5c4-420508603e10-kube-api-access-9h65f\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.252239 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bddeab55-cf79-4073-b5c4-420508603e10-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.280360 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"bddeab55-cf79-4073-b5c4-420508603e10\") " pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.363191 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0b9711c-4309-40aa-8902-e2ad54e47000" path="/var/lib/kubelet/pods/e0b9711c-4309-40aa-8902-e2ad54e47000/volumes" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.502529 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.591058 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-txw9q"] Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.721730 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.862629 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c03d629c-1abf-440f-b3c8-d70e93e36eb1-erlang-cookie-secret\") pod \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.862697 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-config-data\") pod \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.862794 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.862848 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-plugins-conf\") pod \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.862882 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-plugins\") pod \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.862938 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-server-conf\") pod \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.862962 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-tls\") pod \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.863032 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldj6w\" (UniqueName: \"kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-kube-api-access-ldj6w\") pod \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.863066 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c03d629c-1abf-440f-b3c8-d70e93e36eb1-pod-info\") pod \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.863100 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-erlang-cookie\") pod \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.863189 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-confd\") pod \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\" (UID: \"c03d629c-1abf-440f-b3c8-d70e93e36eb1\") " Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.863883 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c03d629c-1abf-440f-b3c8-d70e93e36eb1" (UID: "c03d629c-1abf-440f-b3c8-d70e93e36eb1"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.864147 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.866403 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c03d629c-1abf-440f-b3c8-d70e93e36eb1" (UID: "c03d629c-1abf-440f-b3c8-d70e93e36eb1"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.866437 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c03d629c-1abf-440f-b3c8-d70e93e36eb1" (UID: "c03d629c-1abf-440f-b3c8-d70e93e36eb1"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.871843 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c03d629c-1abf-440f-b3c8-d70e93e36eb1-pod-info" (OuterVolumeSpecName: "pod-info") pod "c03d629c-1abf-440f-b3c8-d70e93e36eb1" (UID: "c03d629c-1abf-440f-b3c8-d70e93e36eb1"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.882976 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-kube-api-access-ldj6w" (OuterVolumeSpecName: "kube-api-access-ldj6w") pod "c03d629c-1abf-440f-b3c8-d70e93e36eb1" (UID: "c03d629c-1abf-440f-b3c8-d70e93e36eb1"). InnerVolumeSpecName "kube-api-access-ldj6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.888521 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03d629c-1abf-440f-b3c8-d70e93e36eb1-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c03d629c-1abf-440f-b3c8-d70e93e36eb1" (UID: "c03d629c-1abf-440f-b3c8-d70e93e36eb1"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.891477 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "c03d629c-1abf-440f-b3c8-d70e93e36eb1" (UID: "c03d629c-1abf-440f-b3c8-d70e93e36eb1"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.894935 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c03d629c-1abf-440f-b3c8-d70e93e36eb1" (UID: "c03d629c-1abf-440f-b3c8-d70e93e36eb1"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.944095 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-config-data" (OuterVolumeSpecName: "config-data") pod "c03d629c-1abf-440f-b3c8-d70e93e36eb1" (UID: "c03d629c-1abf-440f-b3c8-d70e93e36eb1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.970844 4808 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.970881 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.970895 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldj6w\" (UniqueName: \"kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-kube-api-access-ldj6w\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.970911 4808 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c03d629c-1abf-440f-b3c8-d70e93e36eb1-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.970924 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.970936 4808 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c03d629c-1abf-440f-b3c8-d70e93e36eb1-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.970948 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.970974 4808 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.982085 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-server-conf" (OuterVolumeSpecName: "server-conf") pod "c03d629c-1abf-440f-b3c8-d70e93e36eb1" (UID: "c03d629c-1abf-440f-b3c8-d70e93e36eb1"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:46:58 crc kubenswrapper[4808]: I1124 17:46:58.997367 4808 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.003687 4808 generic.go:334] "Generic (PLEG): container finished" podID="c03d629c-1abf-440f-b3c8-d70e93e36eb1" containerID="fc9eaa638279c561b3f3ba4fdf1622a85a985a4623528792cf646cc044e0b0c3" exitCode=0 Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.003739 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c03d629c-1abf-440f-b3c8-d70e93e36eb1","Type":"ContainerDied","Data":"fc9eaa638279c561b3f3ba4fdf1622a85a985a4623528792cf646cc044e0b0c3"} Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.003762 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c03d629c-1abf-440f-b3c8-d70e93e36eb1","Type":"ContainerDied","Data":"19e8dd5c2b161f1c8ec3a27dd487466aeff7ea5927605c36949d21f883884e2c"} Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.003779 4808 scope.go:117] "RemoveContainer" containerID="fc9eaa638279c561b3f3ba4fdf1622a85a985a4623528792cf646cc044e0b0c3" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.003913 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.020170 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" event={"ID":"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab","Type":"ContainerStarted","Data":"20210681ba806116e90255b5e59e5e88ecb1e2b39c8aad8b8b38ada0728c13f0"} Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.037598 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c03d629c-1abf-440f-b3c8-d70e93e36eb1" (UID: "c03d629c-1abf-440f-b3c8-d70e93e36eb1"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.058683 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:46:59 crc kubenswrapper[4808]: W1124 17:46:59.059722 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbddeab55_cf79_4073_b5c4_420508603e10.slice/crio-5711df64b461bf10fd73c14fe97264350972cf6dc2ea3e6ae2a0ba430ed9223c WatchSource:0}: Error finding container 5711df64b461bf10fd73c14fe97264350972cf6dc2ea3e6ae2a0ba430ed9223c: Status 404 returned error can't find the container with id 5711df64b461bf10fd73c14fe97264350972cf6dc2ea3e6ae2a0ba430ed9223c Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.072457 4808 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.072480 4808 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c03d629c-1abf-440f-b3c8-d70e93e36eb1-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.072490 4808 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c03d629c-1abf-440f-b3c8-d70e93e36eb1-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.225740 4808 scope.go:117] "RemoveContainer" containerID="c746e5e9913fa0751076b716717cda0a4d7faa73d4ccd89866e3c4908aef04bf" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.254063 4808 scope.go:117] "RemoveContainer" containerID="fc9eaa638279c561b3f3ba4fdf1622a85a985a4623528792cf646cc044e0b0c3" Nov 24 17:46:59 crc kubenswrapper[4808]: E1124 17:46:59.254619 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc9eaa638279c561b3f3ba4fdf1622a85a985a4623528792cf646cc044e0b0c3\": container with ID starting with fc9eaa638279c561b3f3ba4fdf1622a85a985a4623528792cf646cc044e0b0c3 not found: ID does not exist" containerID="fc9eaa638279c561b3f3ba4fdf1622a85a985a4623528792cf646cc044e0b0c3" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.254661 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc9eaa638279c561b3f3ba4fdf1622a85a985a4623528792cf646cc044e0b0c3"} err="failed to get container status \"fc9eaa638279c561b3f3ba4fdf1622a85a985a4623528792cf646cc044e0b0c3\": rpc error: code = NotFound desc = could not find container \"fc9eaa638279c561b3f3ba4fdf1622a85a985a4623528792cf646cc044e0b0c3\": container with ID starting with fc9eaa638279c561b3f3ba4fdf1622a85a985a4623528792cf646cc044e0b0c3 not found: ID does not exist" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.254686 4808 scope.go:117] "RemoveContainer" containerID="c746e5e9913fa0751076b716717cda0a4d7faa73d4ccd89866e3c4908aef04bf" Nov 24 17:46:59 crc kubenswrapper[4808]: E1124 17:46:59.255047 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c746e5e9913fa0751076b716717cda0a4d7faa73d4ccd89866e3c4908aef04bf\": container with ID starting with c746e5e9913fa0751076b716717cda0a4d7faa73d4ccd89866e3c4908aef04bf not found: ID does not exist" containerID="c746e5e9913fa0751076b716717cda0a4d7faa73d4ccd89866e3c4908aef04bf" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.255073 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c746e5e9913fa0751076b716717cda0a4d7faa73d4ccd89866e3c4908aef04bf"} err="failed to get container status \"c746e5e9913fa0751076b716717cda0a4d7faa73d4ccd89866e3c4908aef04bf\": rpc error: code = NotFound desc = could not find container \"c746e5e9913fa0751076b716717cda0a4d7faa73d4ccd89866e3c4908aef04bf\": container with ID starting with c746e5e9913fa0751076b716717cda0a4d7faa73d4ccd89866e3c4908aef04bf not found: ID does not exist" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.341217 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.354139 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.365600 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:46:59 crc kubenswrapper[4808]: E1124 17:46:59.366040 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c03d629c-1abf-440f-b3c8-d70e93e36eb1" containerName="setup-container" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.366062 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c03d629c-1abf-440f-b3c8-d70e93e36eb1" containerName="setup-container" Nov 24 17:46:59 crc kubenswrapper[4808]: E1124 17:46:59.366095 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c03d629c-1abf-440f-b3c8-d70e93e36eb1" containerName="rabbitmq" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.366103 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c03d629c-1abf-440f-b3c8-d70e93e36eb1" containerName="rabbitmq" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.368719 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c03d629c-1abf-440f-b3c8-d70e93e36eb1" containerName="rabbitmq" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.369906 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.371277 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.372688 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.372855 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.373034 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.373262 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.373554 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-hvrrh" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.373746 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.388860 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.481921 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78791f0f-ea96-45ba-9735-9d10ddebea10-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.482060 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78791f0f-ea96-45ba-9735-9d10ddebea10-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.482086 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.482104 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78791f0f-ea96-45ba-9735-9d10ddebea10-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.482546 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9228\" (UniqueName: \"kubernetes.io/projected/78791f0f-ea96-45ba-9735-9d10ddebea10-kube-api-access-q9228\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.482606 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78791f0f-ea96-45ba-9735-9d10ddebea10-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.482698 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78791f0f-ea96-45ba-9735-9d10ddebea10-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.482962 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78791f0f-ea96-45ba-9735-9d10ddebea10-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.483120 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78791f0f-ea96-45ba-9735-9d10ddebea10-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.483151 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/78791f0f-ea96-45ba-9735-9d10ddebea10-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.483212 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/78791f0f-ea96-45ba-9735-9d10ddebea10-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.585396 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78791f0f-ea96-45ba-9735-9d10ddebea10-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.585477 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78791f0f-ea96-45ba-9735-9d10ddebea10-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.585501 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/78791f0f-ea96-45ba-9735-9d10ddebea10-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.585546 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/78791f0f-ea96-45ba-9735-9d10ddebea10-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.585600 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78791f0f-ea96-45ba-9735-9d10ddebea10-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.585637 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78791f0f-ea96-45ba-9735-9d10ddebea10-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.585657 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.585677 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78791f0f-ea96-45ba-9735-9d10ddebea10-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.585745 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9228\" (UniqueName: \"kubernetes.io/projected/78791f0f-ea96-45ba-9735-9d10ddebea10-kube-api-access-q9228\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.585772 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78791f0f-ea96-45ba-9735-9d10ddebea10-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.585815 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78791f0f-ea96-45ba-9735-9d10ddebea10-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.586225 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78791f0f-ea96-45ba-9735-9d10ddebea10-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.587114 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78791f0f-ea96-45ba-9735-9d10ddebea10-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.587291 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/78791f0f-ea96-45ba-9735-9d10ddebea10-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.587308 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78791f0f-ea96-45ba-9735-9d10ddebea10-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.587446 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.588809 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78791f0f-ea96-45ba-9735-9d10ddebea10-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.590670 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78791f0f-ea96-45ba-9735-9d10ddebea10-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.590853 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78791f0f-ea96-45ba-9735-9d10ddebea10-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.591193 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78791f0f-ea96-45ba-9735-9d10ddebea10-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.592662 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/78791f0f-ea96-45ba-9735-9d10ddebea10-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.606354 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9228\" (UniqueName: \"kubernetes.io/projected/78791f0f-ea96-45ba-9735-9d10ddebea10-kube-api-access-q9228\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.619075 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"78791f0f-ea96-45ba-9735-9d10ddebea10\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:46:59 crc kubenswrapper[4808]: I1124 17:46:59.691326 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:47:00 crc kubenswrapper[4808]: I1124 17:47:00.034511 4808 generic.go:334] "Generic (PLEG): container finished" podID="5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" containerID="097d4fe19824933960cec538f29a623c100dc6867197f218a9dd96665948c57d" exitCode=0 Nov 24 17:47:00 crc kubenswrapper[4808]: I1124 17:47:00.034637 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" event={"ID":"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab","Type":"ContainerDied","Data":"097d4fe19824933960cec538f29a623c100dc6867197f218a9dd96665948c57d"} Nov 24 17:47:00 crc kubenswrapper[4808]: I1124 17:47:00.035524 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bddeab55-cf79-4073-b5c4-420508603e10","Type":"ContainerStarted","Data":"5711df64b461bf10fd73c14fe97264350972cf6dc2ea3e6ae2a0ba430ed9223c"} Nov 24 17:47:00 crc kubenswrapper[4808]: I1124 17:47:00.130599 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:47:00 crc kubenswrapper[4808]: W1124 17:47:00.133462 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78791f0f_ea96_45ba_9735_9d10ddebea10.slice/crio-a18af812f3eb1efc7b9040a66cade6f426f833a4c6cdca786f93ba24ee13c38e WatchSource:0}: Error finding container a18af812f3eb1efc7b9040a66cade6f426f833a4c6cdca786f93ba24ee13c38e: Status 404 returned error can't find the container with id a18af812f3eb1efc7b9040a66cade6f426f833a4c6cdca786f93ba24ee13c38e Nov 24 17:47:00 crc kubenswrapper[4808]: I1124 17:47:00.360599 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03d629c-1abf-440f-b3c8-d70e93e36eb1" path="/var/lib/kubelet/pods/c03d629c-1abf-440f-b3c8-d70e93e36eb1/volumes" Nov 24 17:47:01 crc kubenswrapper[4808]: I1124 17:47:01.071260 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" event={"ID":"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab","Type":"ContainerStarted","Data":"1afe17be703a3172a07b89b94843f4ae83546c4b9dbb2aba15e617563522d64e"} Nov 24 17:47:01 crc kubenswrapper[4808]: I1124 17:47:01.072396 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:47:01 crc kubenswrapper[4808]: I1124 17:47:01.074174 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"78791f0f-ea96-45ba-9735-9d10ddebea10","Type":"ContainerStarted","Data":"a18af812f3eb1efc7b9040a66cade6f426f833a4c6cdca786f93ba24ee13c38e"} Nov 24 17:47:01 crc kubenswrapper[4808]: I1124 17:47:01.076454 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bddeab55-cf79-4073-b5c4-420508603e10","Type":"ContainerStarted","Data":"2a98a633940bd89cf51d773f21f523c54a8decca77a1ca7c7ee788835df79bea"} Nov 24 17:47:01 crc kubenswrapper[4808]: I1124 17:47:01.130194 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" podStartSLOduration=4.130170438 podStartE2EDuration="4.130170438s" podCreationTimestamp="2025-11-24 17:46:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:47:01.097653679 +0000 UTC m=+1213.695321481" watchObservedRunningTime="2025-11-24 17:47:01.130170438 +0000 UTC m=+1213.727838240" Nov 24 17:47:03 crc kubenswrapper[4808]: I1124 17:47:03.096925 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"78791f0f-ea96-45ba-9735-9d10ddebea10","Type":"ContainerStarted","Data":"ddd795c0886392c9bd12d0201623b1165d1df1717f6d8826eb874f87d86a2e3f"} Nov 24 17:47:06 crc kubenswrapper[4808]: I1124 17:47:06.523159 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:47:06 crc kubenswrapper[4808]: I1124 17:47:06.524230 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:47:06 crc kubenswrapper[4808]: I1124 17:47:06.524314 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:47:06 crc kubenswrapper[4808]: I1124 17:47:06.525759 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"38d3a2110cebc7bb322ebd71a0ddbf64043c7ffce95b69f693d9dde312b5d138"} pod="openshift-machine-config-operator/machine-config-daemon-ht87w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:47:06 crc kubenswrapper[4808]: I1124 17:47:06.525860 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" containerID="cri-o://38d3a2110cebc7bb322ebd71a0ddbf64043c7ffce95b69f693d9dde312b5d138" gracePeriod=600 Nov 24 17:47:07 crc kubenswrapper[4808]: I1124 17:47:07.142265 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerID="38d3a2110cebc7bb322ebd71a0ddbf64043c7ffce95b69f693d9dde312b5d138" exitCode=0 Nov 24 17:47:07 crc kubenswrapper[4808]: I1124 17:47:07.142309 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerDied","Data":"38d3a2110cebc7bb322ebd71a0ddbf64043c7ffce95b69f693d9dde312b5d138"} Nov 24 17:47:07 crc kubenswrapper[4808]: I1124 17:47:07.142348 4808 scope.go:117] "RemoveContainer" containerID="216625ce434c33017d87034c7bea8e120016af1d6be08176847569ea1821e3e1" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.077249 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.150205 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-dzsxl"] Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.156389 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"d1973d5d1c793927c6c85a086c241df6363c531bf6676a4e3fab54356b709ffb"} Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.156487 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" podUID="378c3582-e95c-4cd0-9ab9-91057bc816ef" containerName="dnsmasq-dns" containerID="cri-o://fe6ad8a59f01a937082928d3bce3386eb70288ab7a77ab89bc3ad60574cc60ca" gracePeriod=10 Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.290793 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-m2sjz"] Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.292423 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.314782 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-m2sjz"] Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.449249 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.449305 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.449348 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djf2g\" (UniqueName: \"kubernetes.io/projected/db145eef-fe79-49c1-9fd1-570b506b3367-kube-api-access-djf2g\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.449416 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-config\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.449444 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.449518 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-dns-svc\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.449542 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.550912 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-config\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.551397 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.551517 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-dns-svc\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.551548 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.551646 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.551689 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.551795 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djf2g\" (UniqueName: \"kubernetes.io/projected/db145eef-fe79-49c1-9fd1-570b506b3367-kube-api-access-djf2g\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.552859 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-dns-svc\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.552858 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.553095 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.553671 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-config\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.554153 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.554508 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db145eef-fe79-49c1-9fd1-570b506b3367-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.577849 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djf2g\" (UniqueName: \"kubernetes.io/projected/db145eef-fe79-49c1-9fd1-570b506b3367-kube-api-access-djf2g\") pod \"dnsmasq-dns-55478c4467-m2sjz\" (UID: \"db145eef-fe79-49c1-9fd1-570b506b3367\") " pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.649900 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.700607 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.856928 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-dns-swift-storage-0\") pod \"378c3582-e95c-4cd0-9ab9-91057bc816ef\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.857263 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-config\") pod \"378c3582-e95c-4cd0-9ab9-91057bc816ef\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.857355 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-ovsdbserver-sb\") pod \"378c3582-e95c-4cd0-9ab9-91057bc816ef\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.857402 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vckkv\" (UniqueName: \"kubernetes.io/projected/378c3582-e95c-4cd0-9ab9-91057bc816ef-kube-api-access-vckkv\") pod \"378c3582-e95c-4cd0-9ab9-91057bc816ef\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.857437 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-dns-svc\") pod \"378c3582-e95c-4cd0-9ab9-91057bc816ef\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.857462 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-ovsdbserver-nb\") pod \"378c3582-e95c-4cd0-9ab9-91057bc816ef\" (UID: \"378c3582-e95c-4cd0-9ab9-91057bc816ef\") " Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.868817 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/378c3582-e95c-4cd0-9ab9-91057bc816ef-kube-api-access-vckkv" (OuterVolumeSpecName: "kube-api-access-vckkv") pod "378c3582-e95c-4cd0-9ab9-91057bc816ef" (UID: "378c3582-e95c-4cd0-9ab9-91057bc816ef"). InnerVolumeSpecName "kube-api-access-vckkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.931905 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-config" (OuterVolumeSpecName: "config") pod "378c3582-e95c-4cd0-9ab9-91057bc816ef" (UID: "378c3582-e95c-4cd0-9ab9-91057bc816ef"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.954658 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "378c3582-e95c-4cd0-9ab9-91057bc816ef" (UID: "378c3582-e95c-4cd0-9ab9-91057bc816ef"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.957064 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "378c3582-e95c-4cd0-9ab9-91057bc816ef" (UID: "378c3582-e95c-4cd0-9ab9-91057bc816ef"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.960100 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-m2sjz"] Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.960736 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.960771 4808 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.960786 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.960798 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vckkv\" (UniqueName: \"kubernetes.io/projected/378c3582-e95c-4cd0-9ab9-91057bc816ef-kube-api-access-vckkv\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.961051 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "378c3582-e95c-4cd0-9ab9-91057bc816ef" (UID: "378c3582-e95c-4cd0-9ab9-91057bc816ef"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:47:08 crc kubenswrapper[4808]: I1124 17:47:08.962324 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "378c3582-e95c-4cd0-9ab9-91057bc816ef" (UID: "378c3582-e95c-4cd0-9ab9-91057bc816ef"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:47:09 crc kubenswrapper[4808]: I1124 17:47:09.062709 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:09 crc kubenswrapper[4808]: I1124 17:47:09.062736 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/378c3582-e95c-4cd0-9ab9-91057bc816ef-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:09 crc kubenswrapper[4808]: I1124 17:47:09.167681 4808 generic.go:334] "Generic (PLEG): container finished" podID="378c3582-e95c-4cd0-9ab9-91057bc816ef" containerID="fe6ad8a59f01a937082928d3bce3386eb70288ab7a77ab89bc3ad60574cc60ca" exitCode=0 Nov 24 17:47:09 crc kubenswrapper[4808]: I1124 17:47:09.167852 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" Nov 24 17:47:09 crc kubenswrapper[4808]: I1124 17:47:09.169315 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" event={"ID":"378c3582-e95c-4cd0-9ab9-91057bc816ef","Type":"ContainerDied","Data":"fe6ad8a59f01a937082928d3bce3386eb70288ab7a77ab89bc3ad60574cc60ca"} Nov 24 17:47:09 crc kubenswrapper[4808]: I1124 17:47:09.169359 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-dzsxl" event={"ID":"378c3582-e95c-4cd0-9ab9-91057bc816ef","Type":"ContainerDied","Data":"a711eab003e74bd0ac484d5a0b2c4abb8c044db0d1bcaf319b6a2c3d8cb9804b"} Nov 24 17:47:09 crc kubenswrapper[4808]: I1124 17:47:09.169381 4808 scope.go:117] "RemoveContainer" containerID="fe6ad8a59f01a937082928d3bce3386eb70288ab7a77ab89bc3ad60574cc60ca" Nov 24 17:47:09 crc kubenswrapper[4808]: I1124 17:47:09.171141 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-m2sjz" event={"ID":"db145eef-fe79-49c1-9fd1-570b506b3367","Type":"ContainerStarted","Data":"ef33494ddfe33b2f3fda10364907e48bd3cfefa15a2bba69c7ee017221b598d7"} Nov 24 17:47:09 crc kubenswrapper[4808]: I1124 17:47:09.209389 4808 scope.go:117] "RemoveContainer" containerID="fd0e24ce6bb8f63cb56a5b08752d70d0b017ed0e8cb84c4c933a90a00ef9ab9c" Nov 24 17:47:09 crc kubenswrapper[4808]: I1124 17:47:09.214219 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-dzsxl"] Nov 24 17:47:09 crc kubenswrapper[4808]: I1124 17:47:09.242650 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-dzsxl"] Nov 24 17:47:09 crc kubenswrapper[4808]: I1124 17:47:09.242720 4808 scope.go:117] "RemoveContainer" containerID="fe6ad8a59f01a937082928d3bce3386eb70288ab7a77ab89bc3ad60574cc60ca" Nov 24 17:47:09 crc kubenswrapper[4808]: E1124 17:47:09.243940 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe6ad8a59f01a937082928d3bce3386eb70288ab7a77ab89bc3ad60574cc60ca\": container with ID starting with fe6ad8a59f01a937082928d3bce3386eb70288ab7a77ab89bc3ad60574cc60ca not found: ID does not exist" containerID="fe6ad8a59f01a937082928d3bce3386eb70288ab7a77ab89bc3ad60574cc60ca" Nov 24 17:47:09 crc kubenswrapper[4808]: I1124 17:47:09.244122 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe6ad8a59f01a937082928d3bce3386eb70288ab7a77ab89bc3ad60574cc60ca"} err="failed to get container status \"fe6ad8a59f01a937082928d3bce3386eb70288ab7a77ab89bc3ad60574cc60ca\": rpc error: code = NotFound desc = could not find container \"fe6ad8a59f01a937082928d3bce3386eb70288ab7a77ab89bc3ad60574cc60ca\": container with ID starting with fe6ad8a59f01a937082928d3bce3386eb70288ab7a77ab89bc3ad60574cc60ca not found: ID does not exist" Nov 24 17:47:09 crc kubenswrapper[4808]: I1124 17:47:09.244165 4808 scope.go:117] "RemoveContainer" containerID="fd0e24ce6bb8f63cb56a5b08752d70d0b017ed0e8cb84c4c933a90a00ef9ab9c" Nov 24 17:47:09 crc kubenswrapper[4808]: E1124 17:47:09.245939 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd0e24ce6bb8f63cb56a5b08752d70d0b017ed0e8cb84c4c933a90a00ef9ab9c\": container with ID starting with fd0e24ce6bb8f63cb56a5b08752d70d0b017ed0e8cb84c4c933a90a00ef9ab9c not found: ID does not exist" containerID="fd0e24ce6bb8f63cb56a5b08752d70d0b017ed0e8cb84c4c933a90a00ef9ab9c" Nov 24 17:47:09 crc kubenswrapper[4808]: I1124 17:47:09.245986 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd0e24ce6bb8f63cb56a5b08752d70d0b017ed0e8cb84c4c933a90a00ef9ab9c"} err="failed to get container status \"fd0e24ce6bb8f63cb56a5b08752d70d0b017ed0e8cb84c4c933a90a00ef9ab9c\": rpc error: code = NotFound desc = could not find container \"fd0e24ce6bb8f63cb56a5b08752d70d0b017ed0e8cb84c4c933a90a00ef9ab9c\": container with ID starting with fd0e24ce6bb8f63cb56a5b08752d70d0b017ed0e8cb84c4c933a90a00ef9ab9c not found: ID does not exist" Nov 24 17:47:09 crc kubenswrapper[4808]: E1124 17:47:09.404010 4808 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod378c3582_e95c_4cd0_9ab9_91057bc816ef.slice/crio-a711eab003e74bd0ac484d5a0b2c4abb8c044db0d1bcaf319b6a2c3d8cb9804b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb145eef_fe79_49c1_9fd1_570b506b3367.slice/crio-a8626ebb89e7a354e555d200eb493e9a9f8f010f64bb553309aad0367c9bdb0a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod378c3582_e95c_4cd0_9ab9_91057bc816ef.slice\": RecentStats: unable to find data in memory cache]" Nov 24 17:47:10 crc kubenswrapper[4808]: I1124 17:47:10.181893 4808 generic.go:334] "Generic (PLEG): container finished" podID="db145eef-fe79-49c1-9fd1-570b506b3367" containerID="a8626ebb89e7a354e555d200eb493e9a9f8f010f64bb553309aad0367c9bdb0a" exitCode=0 Nov 24 17:47:10 crc kubenswrapper[4808]: I1124 17:47:10.181999 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-m2sjz" event={"ID":"db145eef-fe79-49c1-9fd1-570b506b3367","Type":"ContainerDied","Data":"a8626ebb89e7a354e555d200eb493e9a9f8f010f64bb553309aad0367c9bdb0a"} Nov 24 17:47:10 crc kubenswrapper[4808]: I1124 17:47:10.360798 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="378c3582-e95c-4cd0-9ab9-91057bc816ef" path="/var/lib/kubelet/pods/378c3582-e95c-4cd0-9ab9-91057bc816ef/volumes" Nov 24 17:47:11 crc kubenswrapper[4808]: I1124 17:47:11.191846 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-m2sjz" event={"ID":"db145eef-fe79-49c1-9fd1-570b506b3367","Type":"ContainerStarted","Data":"dc36a64cdd653f5f2f47199b62942eaebc2a60171465c10ec438b273aef58cb4"} Nov 24 17:47:11 crc kubenswrapper[4808]: I1124 17:47:11.192477 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:11 crc kubenswrapper[4808]: I1124 17:47:11.213540 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-m2sjz" podStartSLOduration=3.213519292 podStartE2EDuration="3.213519292s" podCreationTimestamp="2025-11-24 17:47:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:47:11.209069565 +0000 UTC m=+1223.806737367" watchObservedRunningTime="2025-11-24 17:47:11.213519292 +0000 UTC m=+1223.811187094" Nov 24 17:47:18 crc kubenswrapper[4808]: I1124 17:47:18.651348 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-m2sjz" Nov 24 17:47:18 crc kubenswrapper[4808]: I1124 17:47:18.720701 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-txw9q"] Nov 24 17:47:18 crc kubenswrapper[4808]: I1124 17:47:18.720993 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" podUID="5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" containerName="dnsmasq-dns" containerID="cri-o://1afe17be703a3172a07b89b94843f4ae83546c4b9dbb2aba15e617563522d64e" gracePeriod=10 Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.256178 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.262616 4808 generic.go:334] "Generic (PLEG): container finished" podID="5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" containerID="1afe17be703a3172a07b89b94843f4ae83546c4b9dbb2aba15e617563522d64e" exitCode=0 Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.262659 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" event={"ID":"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab","Type":"ContainerDied","Data":"1afe17be703a3172a07b89b94843f4ae83546c4b9dbb2aba15e617563522d64e"} Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.262689 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" event={"ID":"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab","Type":"ContainerDied","Data":"20210681ba806116e90255b5e59e5e88ecb1e2b39c8aad8b8b38ada0728c13f0"} Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.262689 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-txw9q" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.262757 4808 scope.go:117] "RemoveContainer" containerID="1afe17be703a3172a07b89b94843f4ae83546c4b9dbb2aba15e617563522d64e" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.301908 4808 scope.go:117] "RemoveContainer" containerID="097d4fe19824933960cec538f29a623c100dc6867197f218a9dd96665948c57d" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.347192 4808 scope.go:117] "RemoveContainer" containerID="1afe17be703a3172a07b89b94843f4ae83546c4b9dbb2aba15e617563522d64e" Nov 24 17:47:19 crc kubenswrapper[4808]: E1124 17:47:19.349829 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1afe17be703a3172a07b89b94843f4ae83546c4b9dbb2aba15e617563522d64e\": container with ID starting with 1afe17be703a3172a07b89b94843f4ae83546c4b9dbb2aba15e617563522d64e not found: ID does not exist" containerID="1afe17be703a3172a07b89b94843f4ae83546c4b9dbb2aba15e617563522d64e" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.349867 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1afe17be703a3172a07b89b94843f4ae83546c4b9dbb2aba15e617563522d64e"} err="failed to get container status \"1afe17be703a3172a07b89b94843f4ae83546c4b9dbb2aba15e617563522d64e\": rpc error: code = NotFound desc = could not find container \"1afe17be703a3172a07b89b94843f4ae83546c4b9dbb2aba15e617563522d64e\": container with ID starting with 1afe17be703a3172a07b89b94843f4ae83546c4b9dbb2aba15e617563522d64e not found: ID does not exist" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.349890 4808 scope.go:117] "RemoveContainer" containerID="097d4fe19824933960cec538f29a623c100dc6867197f218a9dd96665948c57d" Nov 24 17:47:19 crc kubenswrapper[4808]: E1124 17:47:19.350176 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"097d4fe19824933960cec538f29a623c100dc6867197f218a9dd96665948c57d\": container with ID starting with 097d4fe19824933960cec538f29a623c100dc6867197f218a9dd96665948c57d not found: ID does not exist" containerID="097d4fe19824933960cec538f29a623c100dc6867197f218a9dd96665948c57d" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.350218 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"097d4fe19824933960cec538f29a623c100dc6867197f218a9dd96665948c57d"} err="failed to get container status \"097d4fe19824933960cec538f29a623c100dc6867197f218a9dd96665948c57d\": rpc error: code = NotFound desc = could not find container \"097d4fe19824933960cec538f29a623c100dc6867197f218a9dd96665948c57d\": container with ID starting with 097d4fe19824933960cec538f29a623c100dc6867197f218a9dd96665948c57d not found: ID does not exist" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.366056 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-dns-swift-storage-0\") pod \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.366141 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8thjb\" (UniqueName: \"kubernetes.io/projected/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-kube-api-access-8thjb\") pod \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.366265 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-openstack-edpm-ipam\") pod \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.366367 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-ovsdbserver-nb\") pod \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.366439 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-config\") pod \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.366478 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-ovsdbserver-sb\") pod \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.366554 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-dns-svc\") pod \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\" (UID: \"5bf1b28b-04b3-4883-ab32-ae1d1972b1ab\") " Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.386826 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-kube-api-access-8thjb" (OuterVolumeSpecName: "kube-api-access-8thjb") pod "5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" (UID: "5bf1b28b-04b3-4883-ab32-ae1d1972b1ab"). InnerVolumeSpecName "kube-api-access-8thjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.423597 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" (UID: "5bf1b28b-04b3-4883-ab32-ae1d1972b1ab"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.426827 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" (UID: "5bf1b28b-04b3-4883-ab32-ae1d1972b1ab"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.430497 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" (UID: "5bf1b28b-04b3-4883-ab32-ae1d1972b1ab"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.431954 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" (UID: "5bf1b28b-04b3-4883-ab32-ae1d1972b1ab"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.432898 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-config" (OuterVolumeSpecName: "config") pod "5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" (UID: "5bf1b28b-04b3-4883-ab32-ae1d1972b1ab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.440820 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" (UID: "5bf1b28b-04b3-4883-ab32-ae1d1972b1ab"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.471007 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.471070 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.471082 4808 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.471093 4808 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.471108 4808 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.471125 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8thjb\" (UniqueName: \"kubernetes.io/projected/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-kube-api-access-8thjb\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.471139 4808 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.636173 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-txw9q"] Nov 24 17:47:19 crc kubenswrapper[4808]: I1124 17:47:19.641714 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-txw9q"] Nov 24 17:47:20 crc kubenswrapper[4808]: I1124 17:47:20.359318 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" path="/var/lib/kubelet/pods/5bf1b28b-04b3-4883-ab32-ae1d1972b1ab/volumes" Nov 24 17:47:31 crc kubenswrapper[4808]: I1124 17:47:31.978378 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl"] Nov 24 17:47:31 crc kubenswrapper[4808]: E1124 17:47:31.979345 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" containerName="init" Nov 24 17:47:31 crc kubenswrapper[4808]: I1124 17:47:31.979357 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" containerName="init" Nov 24 17:47:31 crc kubenswrapper[4808]: E1124 17:47:31.979382 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="378c3582-e95c-4cd0-9ab9-91057bc816ef" containerName="dnsmasq-dns" Nov 24 17:47:31 crc kubenswrapper[4808]: I1124 17:47:31.979389 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="378c3582-e95c-4cd0-9ab9-91057bc816ef" containerName="dnsmasq-dns" Nov 24 17:47:31 crc kubenswrapper[4808]: E1124 17:47:31.979407 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="378c3582-e95c-4cd0-9ab9-91057bc816ef" containerName="init" Nov 24 17:47:31 crc kubenswrapper[4808]: I1124 17:47:31.979413 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="378c3582-e95c-4cd0-9ab9-91057bc816ef" containerName="init" Nov 24 17:47:31 crc kubenswrapper[4808]: E1124 17:47:31.979426 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" containerName="dnsmasq-dns" Nov 24 17:47:31 crc kubenswrapper[4808]: I1124 17:47:31.979432 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" containerName="dnsmasq-dns" Nov 24 17:47:31 crc kubenswrapper[4808]: I1124 17:47:31.979585 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="378c3582-e95c-4cd0-9ab9-91057bc816ef" containerName="dnsmasq-dns" Nov 24 17:47:31 crc kubenswrapper[4808]: I1124 17:47:31.979608 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bf1b28b-04b3-4883-ab32-ae1d1972b1ab" containerName="dnsmasq-dns" Nov 24 17:47:31 crc kubenswrapper[4808]: I1124 17:47:31.980317 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:31 crc kubenswrapper[4808]: I1124 17:47:31.982358 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:47:31 crc kubenswrapper[4808]: I1124 17:47:31.984748 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:47:31 crc kubenswrapper[4808]: I1124 17:47:31.984767 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:47:31 crc kubenswrapper[4808]: I1124 17:47:31.984776 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:47:31 crc kubenswrapper[4808]: I1124 17:47:31.998165 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl"] Nov 24 17:47:32 crc kubenswrapper[4808]: I1124 17:47:32.115911 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:32 crc kubenswrapper[4808]: I1124 17:47:32.116063 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:32 crc kubenswrapper[4808]: I1124 17:47:32.116104 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk9dk\" (UniqueName: \"kubernetes.io/projected/a36913c9-4487-4403-9797-8ebacf1304d7-kube-api-access-qk9dk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:32 crc kubenswrapper[4808]: I1124 17:47:32.116126 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:32 crc kubenswrapper[4808]: I1124 17:47:32.217350 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:32 crc kubenswrapper[4808]: I1124 17:47:32.217484 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:32 crc kubenswrapper[4808]: I1124 17:47:32.217534 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk9dk\" (UniqueName: \"kubernetes.io/projected/a36913c9-4487-4403-9797-8ebacf1304d7-kube-api-access-qk9dk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:32 crc kubenswrapper[4808]: I1124 17:47:32.217559 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:32 crc kubenswrapper[4808]: I1124 17:47:32.223622 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:32 crc kubenswrapper[4808]: I1124 17:47:32.225669 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:32 crc kubenswrapper[4808]: I1124 17:47:32.228994 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:32 crc kubenswrapper[4808]: I1124 17:47:32.233250 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk9dk\" (UniqueName: \"kubernetes.io/projected/a36913c9-4487-4403-9797-8ebacf1304d7-kube-api-access-qk9dk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:32 crc kubenswrapper[4808]: I1124 17:47:32.303576 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:32 crc kubenswrapper[4808]: I1124 17:47:32.802822 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl"] Nov 24 17:47:32 crc kubenswrapper[4808]: I1124 17:47:32.811073 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:47:33 crc kubenswrapper[4808]: I1124 17:47:33.427871 4808 generic.go:334] "Generic (PLEG): container finished" podID="bddeab55-cf79-4073-b5c4-420508603e10" containerID="2a98a633940bd89cf51d773f21f523c54a8decca77a1ca7c7ee788835df79bea" exitCode=0 Nov 24 17:47:33 crc kubenswrapper[4808]: I1124 17:47:33.428530 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bddeab55-cf79-4073-b5c4-420508603e10","Type":"ContainerDied","Data":"2a98a633940bd89cf51d773f21f523c54a8decca77a1ca7c7ee788835df79bea"} Nov 24 17:47:33 crc kubenswrapper[4808]: I1124 17:47:33.432618 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" event={"ID":"a36913c9-4487-4403-9797-8ebacf1304d7","Type":"ContainerStarted","Data":"fea308c26fb844a8ab727a4b92061a2b126574ef91252050021ce054ad3b7601"} Nov 24 17:47:34 crc kubenswrapper[4808]: I1124 17:47:34.444414 4808 generic.go:334] "Generic (PLEG): container finished" podID="78791f0f-ea96-45ba-9735-9d10ddebea10" containerID="ddd795c0886392c9bd12d0201623b1165d1df1717f6d8826eb874f87d86a2e3f" exitCode=0 Nov 24 17:47:34 crc kubenswrapper[4808]: I1124 17:47:34.445052 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"78791f0f-ea96-45ba-9735-9d10ddebea10","Type":"ContainerDied","Data":"ddd795c0886392c9bd12d0201623b1165d1df1717f6d8826eb874f87d86a2e3f"} Nov 24 17:47:34 crc kubenswrapper[4808]: I1124 17:47:34.456850 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bddeab55-cf79-4073-b5c4-420508603e10","Type":"ContainerStarted","Data":"aeebe707bbd6ccbdfea6dcef86bc7bd5ee2103715d890704218001700aa72971"} Nov 24 17:47:34 crc kubenswrapper[4808]: I1124 17:47:34.458666 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 17:47:34 crc kubenswrapper[4808]: I1124 17:47:34.497804 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.497783627 podStartE2EDuration="36.497783627s" podCreationTimestamp="2025-11-24 17:46:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:47:34.493932925 +0000 UTC m=+1247.091600747" watchObservedRunningTime="2025-11-24 17:47:34.497783627 +0000 UTC m=+1247.095451429" Nov 24 17:47:36 crc kubenswrapper[4808]: I1124 17:47:36.483311 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"78791f0f-ea96-45ba-9735-9d10ddebea10","Type":"ContainerStarted","Data":"db2b0f03e1c6c814f8692e2830b0ea281231040eaaee512fb9873c5481d94708"} Nov 24 17:47:36 crc kubenswrapper[4808]: I1124 17:47:36.484089 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:47:36 crc kubenswrapper[4808]: I1124 17:47:36.511541 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.511519751 podStartE2EDuration="37.511519751s" podCreationTimestamp="2025-11-24 17:46:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:47:36.506182455 +0000 UTC m=+1249.103850267" watchObservedRunningTime="2025-11-24 17:47:36.511519751 +0000 UTC m=+1249.109187553" Nov 24 17:47:42 crc kubenswrapper[4808]: I1124 17:47:42.537502 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" event={"ID":"a36913c9-4487-4403-9797-8ebacf1304d7","Type":"ContainerStarted","Data":"ef9a01948fa366601b8fa1b25f6c150e358fe37038e46dcbacfd5de525ab7693"} Nov 24 17:47:42 crc kubenswrapper[4808]: I1124 17:47:42.567378 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" podStartSLOduration=2.8968406509999998 podStartE2EDuration="11.56735244s" podCreationTimestamp="2025-11-24 17:47:31 +0000 UTC" firstStartedPulling="2025-11-24 17:47:32.810731863 +0000 UTC m=+1245.408399665" lastFinishedPulling="2025-11-24 17:47:41.481243652 +0000 UTC m=+1254.078911454" observedRunningTime="2025-11-24 17:47:42.552070953 +0000 UTC m=+1255.149738805" watchObservedRunningTime="2025-11-24 17:47:42.56735244 +0000 UTC m=+1255.165020252" Nov 24 17:47:48 crc kubenswrapper[4808]: I1124 17:47:48.510205 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 17:47:49 crc kubenswrapper[4808]: I1124 17:47:49.696149 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:47:52 crc kubenswrapper[4808]: I1124 17:47:52.629300 4808 generic.go:334] "Generic (PLEG): container finished" podID="a36913c9-4487-4403-9797-8ebacf1304d7" containerID="ef9a01948fa366601b8fa1b25f6c150e358fe37038e46dcbacfd5de525ab7693" exitCode=0 Nov 24 17:47:52 crc kubenswrapper[4808]: I1124 17:47:52.629378 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" event={"ID":"a36913c9-4487-4403-9797-8ebacf1304d7","Type":"ContainerDied","Data":"ef9a01948fa366601b8fa1b25f6c150e358fe37038e46dcbacfd5de525ab7693"} Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.029729 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.156095 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-ssh-key\") pod \"a36913c9-4487-4403-9797-8ebacf1304d7\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.156157 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qk9dk\" (UniqueName: \"kubernetes.io/projected/a36913c9-4487-4403-9797-8ebacf1304d7-kube-api-access-qk9dk\") pod \"a36913c9-4487-4403-9797-8ebacf1304d7\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.156256 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-inventory\") pod \"a36913c9-4487-4403-9797-8ebacf1304d7\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.156300 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-repo-setup-combined-ca-bundle\") pod \"a36913c9-4487-4403-9797-8ebacf1304d7\" (UID: \"a36913c9-4487-4403-9797-8ebacf1304d7\") " Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.162799 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a36913c9-4487-4403-9797-8ebacf1304d7-kube-api-access-qk9dk" (OuterVolumeSpecName: "kube-api-access-qk9dk") pod "a36913c9-4487-4403-9797-8ebacf1304d7" (UID: "a36913c9-4487-4403-9797-8ebacf1304d7"). InnerVolumeSpecName "kube-api-access-qk9dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.163080 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "a36913c9-4487-4403-9797-8ebacf1304d7" (UID: "a36913c9-4487-4403-9797-8ebacf1304d7"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.184840 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-inventory" (OuterVolumeSpecName: "inventory") pod "a36913c9-4487-4403-9797-8ebacf1304d7" (UID: "a36913c9-4487-4403-9797-8ebacf1304d7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.193198 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a36913c9-4487-4403-9797-8ebacf1304d7" (UID: "a36913c9-4487-4403-9797-8ebacf1304d7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.258118 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.258163 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qk9dk\" (UniqueName: \"kubernetes.io/projected/a36913c9-4487-4403-9797-8ebacf1304d7-kube-api-access-qk9dk\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.258174 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.258184 4808 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a36913c9-4487-4403-9797-8ebacf1304d7-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.648221 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" event={"ID":"a36913c9-4487-4403-9797-8ebacf1304d7","Type":"ContainerDied","Data":"fea308c26fb844a8ab727a4b92061a2b126574ef91252050021ce054ad3b7601"} Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.648267 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fea308c26fb844a8ab727a4b92061a2b126574ef91252050021ce054ad3b7601" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.648276 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.727215 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7"] Nov 24 17:47:54 crc kubenswrapper[4808]: E1124 17:47:54.727999 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36913c9-4487-4403-9797-8ebacf1304d7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.728073 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36913c9-4487-4403-9797-8ebacf1304d7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.728326 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="a36913c9-4487-4403-9797-8ebacf1304d7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.729199 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.731827 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.731902 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.732086 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.732129 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.751265 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7"] Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.869295 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw67t\" (UniqueName: \"kubernetes.io/projected/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-kube-api-access-jw67t\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkfw7\" (UID: \"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.869340 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkfw7\" (UID: \"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.869413 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkfw7\" (UID: \"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.973400 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw67t\" (UniqueName: \"kubernetes.io/projected/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-kube-api-access-jw67t\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkfw7\" (UID: \"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.973504 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkfw7\" (UID: \"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.973642 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkfw7\" (UID: \"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.979720 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkfw7\" (UID: \"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.982494 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkfw7\" (UID: \"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" Nov 24 17:47:54 crc kubenswrapper[4808]: I1124 17:47:54.993453 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw67t\" (UniqueName: \"kubernetes.io/projected/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-kube-api-access-jw67t\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkfw7\" (UID: \"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" Nov 24 17:47:55 crc kubenswrapper[4808]: I1124 17:47:55.071894 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" Nov 24 17:47:55 crc kubenswrapper[4808]: I1124 17:47:55.572705 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7"] Nov 24 17:47:55 crc kubenswrapper[4808]: W1124 17:47:55.582208 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27ee00de_fd0b_435d_84a5_28cf3c4d0dc7.slice/crio-d399e8a46fb4c1fa6c86f7dee4eab4d3597df635d4df9158dda0e5e1965e9581 WatchSource:0}: Error finding container d399e8a46fb4c1fa6c86f7dee4eab4d3597df635d4df9158dda0e5e1965e9581: Status 404 returned error can't find the container with id d399e8a46fb4c1fa6c86f7dee4eab4d3597df635d4df9158dda0e5e1965e9581 Nov 24 17:47:55 crc kubenswrapper[4808]: I1124 17:47:55.659230 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" event={"ID":"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7","Type":"ContainerStarted","Data":"d399e8a46fb4c1fa6c86f7dee4eab4d3597df635d4df9158dda0e5e1965e9581"} Nov 24 17:47:56 crc kubenswrapper[4808]: I1124 17:47:56.670794 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" event={"ID":"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7","Type":"ContainerStarted","Data":"ae131bccf74e4fd8b14cebffd7b00c117b2235fe7f385856250b844cc5c946ec"} Nov 24 17:47:56 crc kubenswrapper[4808]: I1124 17:47:56.693571 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" podStartSLOduration=2.275489445 podStartE2EDuration="2.693542155s" podCreationTimestamp="2025-11-24 17:47:54 +0000 UTC" firstStartedPulling="2025-11-24 17:47:55.584838047 +0000 UTC m=+1268.182505849" lastFinishedPulling="2025-11-24 17:47:56.002890757 +0000 UTC m=+1268.600558559" observedRunningTime="2025-11-24 17:47:56.686944652 +0000 UTC m=+1269.284612474" watchObservedRunningTime="2025-11-24 17:47:56.693542155 +0000 UTC m=+1269.291209977" Nov 24 17:47:58 crc kubenswrapper[4808]: I1124 17:47:58.705459 4808 generic.go:334] "Generic (PLEG): container finished" podID="27ee00de-fd0b-435d-84a5-28cf3c4d0dc7" containerID="ae131bccf74e4fd8b14cebffd7b00c117b2235fe7f385856250b844cc5c946ec" exitCode=0 Nov 24 17:47:58 crc kubenswrapper[4808]: I1124 17:47:58.705585 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" event={"ID":"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7","Type":"ContainerDied","Data":"ae131bccf74e4fd8b14cebffd7b00c117b2235fe7f385856250b844cc5c946ec"} Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.104108 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.166041 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-inventory\") pod \"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7\" (UID: \"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7\") " Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.166103 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-ssh-key\") pod \"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7\" (UID: \"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7\") " Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.166274 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jw67t\" (UniqueName: \"kubernetes.io/projected/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-kube-api-access-jw67t\") pod \"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7\" (UID: \"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7\") " Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.172189 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-kube-api-access-jw67t" (OuterVolumeSpecName: "kube-api-access-jw67t") pod "27ee00de-fd0b-435d-84a5-28cf3c4d0dc7" (UID: "27ee00de-fd0b-435d-84a5-28cf3c4d0dc7"). InnerVolumeSpecName "kube-api-access-jw67t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.194345 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "27ee00de-fd0b-435d-84a5-28cf3c4d0dc7" (UID: "27ee00de-fd0b-435d-84a5-28cf3c4d0dc7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.194503 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-inventory" (OuterVolumeSpecName: "inventory") pod "27ee00de-fd0b-435d-84a5-28cf3c4d0dc7" (UID: "27ee00de-fd0b-435d-84a5-28cf3c4d0dc7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.268455 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.268483 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.268492 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jw67t\" (UniqueName: \"kubernetes.io/projected/27ee00de-fd0b-435d-84a5-28cf3c4d0dc7-kube-api-access-jw67t\") on node \"crc\" DevicePath \"\"" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.725164 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" event={"ID":"27ee00de-fd0b-435d-84a5-28cf3c4d0dc7","Type":"ContainerDied","Data":"d399e8a46fb4c1fa6c86f7dee4eab4d3597df635d4df9158dda0e5e1965e9581"} Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.725204 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d399e8a46fb4c1fa6c86f7dee4eab4d3597df635d4df9158dda0e5e1965e9581" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.725224 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkfw7" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.786435 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr"] Nov 24 17:48:00 crc kubenswrapper[4808]: E1124 17:48:00.786773 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27ee00de-fd0b-435d-84a5-28cf3c4d0dc7" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.786790 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="27ee00de-fd0b-435d-84a5-28cf3c4d0dc7" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.786990 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="27ee00de-fd0b-435d-84a5-28cf3c4d0dc7" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.787542 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.789623 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.789651 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.789850 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.789999 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.820815 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr"] Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.878871 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkkdq\" (UniqueName: \"kubernetes.io/projected/b6ec6116-79f6-4178-9ebc-bc29c07889b3-kube-api-access-nkkdq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.879146 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.879298 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.879359 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.980937 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.981413 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkkdq\" (UniqueName: \"kubernetes.io/projected/b6ec6116-79f6-4178-9ebc-bc29c07889b3-kube-api-access-nkkdq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.981513 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.981638 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.987066 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.987716 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:48:00 crc kubenswrapper[4808]: I1124 17:48:00.988060 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:48:01 crc kubenswrapper[4808]: I1124 17:48:01.000507 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkkdq\" (UniqueName: \"kubernetes.io/projected/b6ec6116-79f6-4178-9ebc-bc29c07889b3-kube-api-access-nkkdq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:48:01 crc kubenswrapper[4808]: I1124 17:48:01.115156 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:48:01 crc kubenswrapper[4808]: I1124 17:48:01.698519 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr"] Nov 24 17:48:01 crc kubenswrapper[4808]: I1124 17:48:01.734543 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" event={"ID":"b6ec6116-79f6-4178-9ebc-bc29c07889b3","Type":"ContainerStarted","Data":"e36e94e3708a372fc24d1ce415cb58793fd1823e8d08d0e569236cdd54ed31cc"} Nov 24 17:48:02 crc kubenswrapper[4808]: I1124 17:48:02.745233 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" event={"ID":"b6ec6116-79f6-4178-9ebc-bc29c07889b3","Type":"ContainerStarted","Data":"4a2b93b0c0722322a99cc63b1c97a9fac25ca1f5e13ff64e848a3d7e58830e85"} Nov 24 17:48:02 crc kubenswrapper[4808]: I1124 17:48:02.771050 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" podStartSLOduration=2.006431017 podStartE2EDuration="2.771027547s" podCreationTimestamp="2025-11-24 17:48:00 +0000 UTC" firstStartedPulling="2025-11-24 17:48:01.713314449 +0000 UTC m=+1274.310982271" lastFinishedPulling="2025-11-24 17:48:02.477910979 +0000 UTC m=+1275.075578801" observedRunningTime="2025-11-24 17:48:02.766552956 +0000 UTC m=+1275.364220768" watchObservedRunningTime="2025-11-24 17:48:02.771027547 +0000 UTC m=+1275.368695349" Nov 24 17:48:52 crc kubenswrapper[4808]: I1124 17:48:52.149060 4808 scope.go:117] "RemoveContainer" containerID="9b40e76510974e9bec17fedccedaa306a188f9d7a1f0451ec0d8474a6d752e1c" Nov 24 17:48:52 crc kubenswrapper[4808]: I1124 17:48:52.177174 4808 scope.go:117] "RemoveContainer" containerID="d04912d83a47d1161dbc73f3b5b80134fa4d250a92b7071f0aef10f522bf0dbb" Nov 24 17:49:36 crc kubenswrapper[4808]: I1124 17:49:36.522737 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:49:36 crc kubenswrapper[4808]: I1124 17:49:36.523426 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:49:52 crc kubenswrapper[4808]: I1124 17:49:52.281829 4808 scope.go:117] "RemoveContainer" containerID="9bd29801d8ef589fdf3b93006e8a0418eafd525e3f324cd7540d1c30ee458604" Nov 24 17:49:52 crc kubenswrapper[4808]: I1124 17:49:52.303964 4808 scope.go:117] "RemoveContainer" containerID="2ae865cf786b6e42f1815bc5e71fb46309b6c2b051521e8390781bc858facd93" Nov 24 17:49:52 crc kubenswrapper[4808]: I1124 17:49:52.329774 4808 scope.go:117] "RemoveContainer" containerID="fbc61c0dd82b553d021d24e75615ff8ba915ee6ff26a60b71262cac688eabbcf" Nov 24 17:50:06 crc kubenswrapper[4808]: I1124 17:50:06.523187 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:50:06 crc kubenswrapper[4808]: I1124 17:50:06.524334 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:50:08 crc kubenswrapper[4808]: I1124 17:50:08.963697 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5znqt"] Nov 24 17:50:08 crc kubenswrapper[4808]: I1124 17:50:08.966306 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:08 crc kubenswrapper[4808]: I1124 17:50:08.975981 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5znqt"] Nov 24 17:50:09 crc kubenswrapper[4808]: I1124 17:50:09.108729 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d33b7f5e-7095-4eaa-baf8-339d235e88ba-catalog-content\") pod \"community-operators-5znqt\" (UID: \"d33b7f5e-7095-4eaa-baf8-339d235e88ba\") " pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:09 crc kubenswrapper[4808]: I1124 17:50:09.108826 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svjc7\" (UniqueName: \"kubernetes.io/projected/d33b7f5e-7095-4eaa-baf8-339d235e88ba-kube-api-access-svjc7\") pod \"community-operators-5znqt\" (UID: \"d33b7f5e-7095-4eaa-baf8-339d235e88ba\") " pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:09 crc kubenswrapper[4808]: I1124 17:50:09.108879 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d33b7f5e-7095-4eaa-baf8-339d235e88ba-utilities\") pod \"community-operators-5znqt\" (UID: \"d33b7f5e-7095-4eaa-baf8-339d235e88ba\") " pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:09 crc kubenswrapper[4808]: I1124 17:50:09.210715 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d33b7f5e-7095-4eaa-baf8-339d235e88ba-catalog-content\") pod \"community-operators-5znqt\" (UID: \"d33b7f5e-7095-4eaa-baf8-339d235e88ba\") " pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:09 crc kubenswrapper[4808]: I1124 17:50:09.210806 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svjc7\" (UniqueName: \"kubernetes.io/projected/d33b7f5e-7095-4eaa-baf8-339d235e88ba-kube-api-access-svjc7\") pod \"community-operators-5znqt\" (UID: \"d33b7f5e-7095-4eaa-baf8-339d235e88ba\") " pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:09 crc kubenswrapper[4808]: I1124 17:50:09.210849 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d33b7f5e-7095-4eaa-baf8-339d235e88ba-utilities\") pod \"community-operators-5znqt\" (UID: \"d33b7f5e-7095-4eaa-baf8-339d235e88ba\") " pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:09 crc kubenswrapper[4808]: I1124 17:50:09.211433 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d33b7f5e-7095-4eaa-baf8-339d235e88ba-catalog-content\") pod \"community-operators-5znqt\" (UID: \"d33b7f5e-7095-4eaa-baf8-339d235e88ba\") " pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:09 crc kubenswrapper[4808]: I1124 17:50:09.211713 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d33b7f5e-7095-4eaa-baf8-339d235e88ba-utilities\") pod \"community-operators-5znqt\" (UID: \"d33b7f5e-7095-4eaa-baf8-339d235e88ba\") " pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:09 crc kubenswrapper[4808]: I1124 17:50:09.230856 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svjc7\" (UniqueName: \"kubernetes.io/projected/d33b7f5e-7095-4eaa-baf8-339d235e88ba-kube-api-access-svjc7\") pod \"community-operators-5znqt\" (UID: \"d33b7f5e-7095-4eaa-baf8-339d235e88ba\") " pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:09 crc kubenswrapper[4808]: I1124 17:50:09.298585 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:10 crc kubenswrapper[4808]: I1124 17:50:10.045360 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5znqt"] Nov 24 17:50:10 crc kubenswrapper[4808]: I1124 17:50:10.917286 4808 generic.go:334] "Generic (PLEG): container finished" podID="d33b7f5e-7095-4eaa-baf8-339d235e88ba" containerID="a888247e4026efd28558da6167a3e95cd5cb7bba76aa2f17d716d72edbe93177" exitCode=0 Nov 24 17:50:10 crc kubenswrapper[4808]: I1124 17:50:10.917447 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5znqt" event={"ID":"d33b7f5e-7095-4eaa-baf8-339d235e88ba","Type":"ContainerDied","Data":"a888247e4026efd28558da6167a3e95cd5cb7bba76aa2f17d716d72edbe93177"} Nov 24 17:50:10 crc kubenswrapper[4808]: I1124 17:50:10.917802 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5znqt" event={"ID":"d33b7f5e-7095-4eaa-baf8-339d235e88ba","Type":"ContainerStarted","Data":"5ff25561d6b2281d16dd007f8368ce1d87c71d0f5bfff714fb08241366fc7b73"} Nov 24 17:50:11 crc kubenswrapper[4808]: I1124 17:50:11.929372 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5znqt" event={"ID":"d33b7f5e-7095-4eaa-baf8-339d235e88ba","Type":"ContainerStarted","Data":"c995cab6419d209284f50f4c449ef3773dab3d490c844f78fc2495c3f5902e14"} Nov 24 17:50:12 crc kubenswrapper[4808]: I1124 17:50:12.940613 4808 generic.go:334] "Generic (PLEG): container finished" podID="d33b7f5e-7095-4eaa-baf8-339d235e88ba" containerID="c995cab6419d209284f50f4c449ef3773dab3d490c844f78fc2495c3f5902e14" exitCode=0 Nov 24 17:50:12 crc kubenswrapper[4808]: I1124 17:50:12.940660 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5znqt" event={"ID":"d33b7f5e-7095-4eaa-baf8-339d235e88ba","Type":"ContainerDied","Data":"c995cab6419d209284f50f4c449ef3773dab3d490c844f78fc2495c3f5902e14"} Nov 24 17:50:13 crc kubenswrapper[4808]: I1124 17:50:13.950931 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5znqt" event={"ID":"d33b7f5e-7095-4eaa-baf8-339d235e88ba","Type":"ContainerStarted","Data":"5db6777c3e0c0a7c248c78b9b60d8e3edb23a44ac6148e1c7355e8757ad5dd08"} Nov 24 17:50:13 crc kubenswrapper[4808]: I1124 17:50:13.971965 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5znqt" podStartSLOduration=3.5063745219999998 podStartE2EDuration="5.971948005s" podCreationTimestamp="2025-11-24 17:50:08 +0000 UTC" firstStartedPulling="2025-11-24 17:50:10.919984427 +0000 UTC m=+1403.517652229" lastFinishedPulling="2025-11-24 17:50:13.38555792 +0000 UTC m=+1405.983225712" observedRunningTime="2025-11-24 17:50:13.966495782 +0000 UTC m=+1406.564163594" watchObservedRunningTime="2025-11-24 17:50:13.971948005 +0000 UTC m=+1406.569615807" Nov 24 17:50:19 crc kubenswrapper[4808]: I1124 17:50:19.299392 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:19 crc kubenswrapper[4808]: I1124 17:50:19.299916 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:19 crc kubenswrapper[4808]: I1124 17:50:19.341444 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:20 crc kubenswrapper[4808]: I1124 17:50:20.066967 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:20 crc kubenswrapper[4808]: I1124 17:50:20.106837 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5znqt"] Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.042041 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5znqt" podUID="d33b7f5e-7095-4eaa-baf8-339d235e88ba" containerName="registry-server" containerID="cri-o://5db6777c3e0c0a7c248c78b9b60d8e3edb23a44ac6148e1c7355e8757ad5dd08" gracePeriod=2 Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.058684 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qn9nt"] Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.065333 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.081634 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee37c403-358a-43db-912e-b04670d87178-utilities\") pod \"certified-operators-qn9nt\" (UID: \"ee37c403-358a-43db-912e-b04670d87178\") " pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.081746 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjhwf\" (UniqueName: \"kubernetes.io/projected/ee37c403-358a-43db-912e-b04670d87178-kube-api-access-jjhwf\") pod \"certified-operators-qn9nt\" (UID: \"ee37c403-358a-43db-912e-b04670d87178\") " pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.081851 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee37c403-358a-43db-912e-b04670d87178-catalog-content\") pod \"certified-operators-qn9nt\" (UID: \"ee37c403-358a-43db-912e-b04670d87178\") " pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.102960 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qn9nt"] Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.184584 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee37c403-358a-43db-912e-b04670d87178-catalog-content\") pod \"certified-operators-qn9nt\" (UID: \"ee37c403-358a-43db-912e-b04670d87178\") " pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.184824 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee37c403-358a-43db-912e-b04670d87178-utilities\") pod \"certified-operators-qn9nt\" (UID: \"ee37c403-358a-43db-912e-b04670d87178\") " pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.184878 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjhwf\" (UniqueName: \"kubernetes.io/projected/ee37c403-358a-43db-912e-b04670d87178-kube-api-access-jjhwf\") pod \"certified-operators-qn9nt\" (UID: \"ee37c403-358a-43db-912e-b04670d87178\") " pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.185073 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee37c403-358a-43db-912e-b04670d87178-catalog-content\") pod \"certified-operators-qn9nt\" (UID: \"ee37c403-358a-43db-912e-b04670d87178\") " pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.185273 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee37c403-358a-43db-912e-b04670d87178-utilities\") pod \"certified-operators-qn9nt\" (UID: \"ee37c403-358a-43db-912e-b04670d87178\") " pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.208240 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjhwf\" (UniqueName: \"kubernetes.io/projected/ee37c403-358a-43db-912e-b04670d87178-kube-api-access-jjhwf\") pod \"certified-operators-qn9nt\" (UID: \"ee37c403-358a-43db-912e-b04670d87178\") " pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.389786 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.614663 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.715925 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d33b7f5e-7095-4eaa-baf8-339d235e88ba-catalog-content\") pod \"d33b7f5e-7095-4eaa-baf8-339d235e88ba\" (UID: \"d33b7f5e-7095-4eaa-baf8-339d235e88ba\") " Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.716076 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svjc7\" (UniqueName: \"kubernetes.io/projected/d33b7f5e-7095-4eaa-baf8-339d235e88ba-kube-api-access-svjc7\") pod \"d33b7f5e-7095-4eaa-baf8-339d235e88ba\" (UID: \"d33b7f5e-7095-4eaa-baf8-339d235e88ba\") " Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.716151 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d33b7f5e-7095-4eaa-baf8-339d235e88ba-utilities\") pod \"d33b7f5e-7095-4eaa-baf8-339d235e88ba\" (UID: \"d33b7f5e-7095-4eaa-baf8-339d235e88ba\") " Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.718695 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d33b7f5e-7095-4eaa-baf8-339d235e88ba-utilities" (OuterVolumeSpecName: "utilities") pod "d33b7f5e-7095-4eaa-baf8-339d235e88ba" (UID: "d33b7f5e-7095-4eaa-baf8-339d235e88ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.735355 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d33b7f5e-7095-4eaa-baf8-339d235e88ba-kube-api-access-svjc7" (OuterVolumeSpecName: "kube-api-access-svjc7") pod "d33b7f5e-7095-4eaa-baf8-339d235e88ba" (UID: "d33b7f5e-7095-4eaa-baf8-339d235e88ba"). InnerVolumeSpecName "kube-api-access-svjc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.818694 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svjc7\" (UniqueName: \"kubernetes.io/projected/d33b7f5e-7095-4eaa-baf8-339d235e88ba-kube-api-access-svjc7\") on node \"crc\" DevicePath \"\"" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.818727 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d33b7f5e-7095-4eaa-baf8-339d235e88ba-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:50:22 crc kubenswrapper[4808]: I1124 17:50:22.980228 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qn9nt"] Nov 24 17:50:23 crc kubenswrapper[4808]: I1124 17:50:23.069851 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qn9nt" event={"ID":"ee37c403-358a-43db-912e-b04670d87178","Type":"ContainerStarted","Data":"d201dad0935d1fb58609c6a1fade8f9b79246d0433b9a920027110caa20bb7e1"} Nov 24 17:50:23 crc kubenswrapper[4808]: I1124 17:50:23.072550 4808 generic.go:334] "Generic (PLEG): container finished" podID="d33b7f5e-7095-4eaa-baf8-339d235e88ba" containerID="5db6777c3e0c0a7c248c78b9b60d8e3edb23a44ac6148e1c7355e8757ad5dd08" exitCode=0 Nov 24 17:50:23 crc kubenswrapper[4808]: I1124 17:50:23.072599 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5znqt" event={"ID":"d33b7f5e-7095-4eaa-baf8-339d235e88ba","Type":"ContainerDied","Data":"5db6777c3e0c0a7c248c78b9b60d8e3edb23a44ac6148e1c7355e8757ad5dd08"} Nov 24 17:50:23 crc kubenswrapper[4808]: I1124 17:50:23.072622 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5znqt" event={"ID":"d33b7f5e-7095-4eaa-baf8-339d235e88ba","Type":"ContainerDied","Data":"5ff25561d6b2281d16dd007f8368ce1d87c71d0f5bfff714fb08241366fc7b73"} Nov 24 17:50:23 crc kubenswrapper[4808]: I1124 17:50:23.072638 4808 scope.go:117] "RemoveContainer" containerID="5db6777c3e0c0a7c248c78b9b60d8e3edb23a44ac6148e1c7355e8757ad5dd08" Nov 24 17:50:23 crc kubenswrapper[4808]: I1124 17:50:23.072680 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5znqt" Nov 24 17:50:23 crc kubenswrapper[4808]: I1124 17:50:23.105866 4808 scope.go:117] "RemoveContainer" containerID="c995cab6419d209284f50f4c449ef3773dab3d490c844f78fc2495c3f5902e14" Nov 24 17:50:23 crc kubenswrapper[4808]: I1124 17:50:23.175100 4808 scope.go:117] "RemoveContainer" containerID="a888247e4026efd28558da6167a3e95cd5cb7bba76aa2f17d716d72edbe93177" Nov 24 17:50:23 crc kubenswrapper[4808]: I1124 17:50:23.196096 4808 scope.go:117] "RemoveContainer" containerID="5db6777c3e0c0a7c248c78b9b60d8e3edb23a44ac6148e1c7355e8757ad5dd08" Nov 24 17:50:23 crc kubenswrapper[4808]: E1124 17:50:23.196449 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5db6777c3e0c0a7c248c78b9b60d8e3edb23a44ac6148e1c7355e8757ad5dd08\": container with ID starting with 5db6777c3e0c0a7c248c78b9b60d8e3edb23a44ac6148e1c7355e8757ad5dd08 not found: ID does not exist" containerID="5db6777c3e0c0a7c248c78b9b60d8e3edb23a44ac6148e1c7355e8757ad5dd08" Nov 24 17:50:23 crc kubenswrapper[4808]: I1124 17:50:23.196499 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5db6777c3e0c0a7c248c78b9b60d8e3edb23a44ac6148e1c7355e8757ad5dd08"} err="failed to get container status \"5db6777c3e0c0a7c248c78b9b60d8e3edb23a44ac6148e1c7355e8757ad5dd08\": rpc error: code = NotFound desc = could not find container \"5db6777c3e0c0a7c248c78b9b60d8e3edb23a44ac6148e1c7355e8757ad5dd08\": container with ID starting with 5db6777c3e0c0a7c248c78b9b60d8e3edb23a44ac6148e1c7355e8757ad5dd08 not found: ID does not exist" Nov 24 17:50:23 crc kubenswrapper[4808]: I1124 17:50:23.196526 4808 scope.go:117] "RemoveContainer" containerID="c995cab6419d209284f50f4c449ef3773dab3d490c844f78fc2495c3f5902e14" Nov 24 17:50:23 crc kubenswrapper[4808]: E1124 17:50:23.196787 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c995cab6419d209284f50f4c449ef3773dab3d490c844f78fc2495c3f5902e14\": container with ID starting with c995cab6419d209284f50f4c449ef3773dab3d490c844f78fc2495c3f5902e14 not found: ID does not exist" containerID="c995cab6419d209284f50f4c449ef3773dab3d490c844f78fc2495c3f5902e14" Nov 24 17:50:23 crc kubenswrapper[4808]: I1124 17:50:23.196815 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c995cab6419d209284f50f4c449ef3773dab3d490c844f78fc2495c3f5902e14"} err="failed to get container status \"c995cab6419d209284f50f4c449ef3773dab3d490c844f78fc2495c3f5902e14\": rpc error: code = NotFound desc = could not find container \"c995cab6419d209284f50f4c449ef3773dab3d490c844f78fc2495c3f5902e14\": container with ID starting with c995cab6419d209284f50f4c449ef3773dab3d490c844f78fc2495c3f5902e14 not found: ID does not exist" Nov 24 17:50:23 crc kubenswrapper[4808]: I1124 17:50:23.196833 4808 scope.go:117] "RemoveContainer" containerID="a888247e4026efd28558da6167a3e95cd5cb7bba76aa2f17d716d72edbe93177" Nov 24 17:50:23 crc kubenswrapper[4808]: E1124 17:50:23.197082 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a888247e4026efd28558da6167a3e95cd5cb7bba76aa2f17d716d72edbe93177\": container with ID starting with a888247e4026efd28558da6167a3e95cd5cb7bba76aa2f17d716d72edbe93177 not found: ID does not exist" containerID="a888247e4026efd28558da6167a3e95cd5cb7bba76aa2f17d716d72edbe93177" Nov 24 17:50:23 crc kubenswrapper[4808]: I1124 17:50:23.197100 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a888247e4026efd28558da6167a3e95cd5cb7bba76aa2f17d716d72edbe93177"} err="failed to get container status \"a888247e4026efd28558da6167a3e95cd5cb7bba76aa2f17d716d72edbe93177\": rpc error: code = NotFound desc = could not find container \"a888247e4026efd28558da6167a3e95cd5cb7bba76aa2f17d716d72edbe93177\": container with ID starting with a888247e4026efd28558da6167a3e95cd5cb7bba76aa2f17d716d72edbe93177 not found: ID does not exist" Nov 24 17:50:24 crc kubenswrapper[4808]: E1124 17:50:24.021105 4808 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee37c403_358a_43db_912e_b04670d87178.slice/crio-conmon-63fb509837d612a84c7c247347a8c385822ba2ce8bef63384f0a733488c6a4d4.scope\": RecentStats: unable to find data in memory cache]" Nov 24 17:50:24 crc kubenswrapper[4808]: I1124 17:50:24.086809 4808 generic.go:334] "Generic (PLEG): container finished" podID="ee37c403-358a-43db-912e-b04670d87178" containerID="63fb509837d612a84c7c247347a8c385822ba2ce8bef63384f0a733488c6a4d4" exitCode=0 Nov 24 17:50:24 crc kubenswrapper[4808]: I1124 17:50:24.086861 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qn9nt" event={"ID":"ee37c403-358a-43db-912e-b04670d87178","Type":"ContainerDied","Data":"63fb509837d612a84c7c247347a8c385822ba2ce8bef63384f0a733488c6a4d4"} Nov 24 17:50:24 crc kubenswrapper[4808]: I1124 17:50:24.276538 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d33b7f5e-7095-4eaa-baf8-339d235e88ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d33b7f5e-7095-4eaa-baf8-339d235e88ba" (UID: "d33b7f5e-7095-4eaa-baf8-339d235e88ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:50:24 crc kubenswrapper[4808]: I1124 17:50:24.311552 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d33b7f5e-7095-4eaa-baf8-339d235e88ba-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:50:24 crc kubenswrapper[4808]: I1124 17:50:24.593500 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5znqt"] Nov 24 17:50:24 crc kubenswrapper[4808]: I1124 17:50:24.601998 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5znqt"] Nov 24 17:50:25 crc kubenswrapper[4808]: I1124 17:50:25.098134 4808 generic.go:334] "Generic (PLEG): container finished" podID="ee37c403-358a-43db-912e-b04670d87178" containerID="56074b5ea32d75ee4c355218d4303ea163ed1b2b286ed9f1b1c80df11ace495a" exitCode=0 Nov 24 17:50:25 crc kubenswrapper[4808]: I1124 17:50:25.098188 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qn9nt" event={"ID":"ee37c403-358a-43db-912e-b04670d87178","Type":"ContainerDied","Data":"56074b5ea32d75ee4c355218d4303ea163ed1b2b286ed9f1b1c80df11ace495a"} Nov 24 17:50:26 crc kubenswrapper[4808]: I1124 17:50:26.109386 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qn9nt" event={"ID":"ee37c403-358a-43db-912e-b04670d87178","Type":"ContainerStarted","Data":"da82936ec4da64c2709a3c1b6ca533dab8e307c442d6861f6ef37b52ec136e7f"} Nov 24 17:50:26 crc kubenswrapper[4808]: I1124 17:50:26.132953 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qn9nt" podStartSLOduration=2.7451091180000002 podStartE2EDuration="4.132933018s" podCreationTimestamp="2025-11-24 17:50:22 +0000 UTC" firstStartedPulling="2025-11-24 17:50:24.090180332 +0000 UTC m=+1416.687848134" lastFinishedPulling="2025-11-24 17:50:25.478004232 +0000 UTC m=+1418.075672034" observedRunningTime="2025-11-24 17:50:26.125109342 +0000 UTC m=+1418.722777174" watchObservedRunningTime="2025-11-24 17:50:26.132933018 +0000 UTC m=+1418.730600820" Nov 24 17:50:26 crc kubenswrapper[4808]: I1124 17:50:26.357770 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d33b7f5e-7095-4eaa-baf8-339d235e88ba" path="/var/lib/kubelet/pods/d33b7f5e-7095-4eaa-baf8-339d235e88ba/volumes" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.594344 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7q7pn"] Nov 24 17:50:29 crc kubenswrapper[4808]: E1124 17:50:29.595368 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d33b7f5e-7095-4eaa-baf8-339d235e88ba" containerName="registry-server" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.595381 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d33b7f5e-7095-4eaa-baf8-339d235e88ba" containerName="registry-server" Nov 24 17:50:29 crc kubenswrapper[4808]: E1124 17:50:29.595402 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d33b7f5e-7095-4eaa-baf8-339d235e88ba" containerName="extract-content" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.595410 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d33b7f5e-7095-4eaa-baf8-339d235e88ba" containerName="extract-content" Nov 24 17:50:29 crc kubenswrapper[4808]: E1124 17:50:29.595426 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d33b7f5e-7095-4eaa-baf8-339d235e88ba" containerName="extract-utilities" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.595433 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d33b7f5e-7095-4eaa-baf8-339d235e88ba" containerName="extract-utilities" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.595626 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="d33b7f5e-7095-4eaa-baf8-339d235e88ba" containerName="registry-server" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.597435 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.607692 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7q7pn"] Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.742278 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17a6501d-7484-4265-9d97-f6c0b4582cf8-catalog-content\") pod \"redhat-operators-7q7pn\" (UID: \"17a6501d-7484-4265-9d97-f6c0b4582cf8\") " pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.742656 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgnvp\" (UniqueName: \"kubernetes.io/projected/17a6501d-7484-4265-9d97-f6c0b4582cf8-kube-api-access-pgnvp\") pod \"redhat-operators-7q7pn\" (UID: \"17a6501d-7484-4265-9d97-f6c0b4582cf8\") " pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.742814 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17a6501d-7484-4265-9d97-f6c0b4582cf8-utilities\") pod \"redhat-operators-7q7pn\" (UID: \"17a6501d-7484-4265-9d97-f6c0b4582cf8\") " pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.844835 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgnvp\" (UniqueName: \"kubernetes.io/projected/17a6501d-7484-4265-9d97-f6c0b4582cf8-kube-api-access-pgnvp\") pod \"redhat-operators-7q7pn\" (UID: \"17a6501d-7484-4265-9d97-f6c0b4582cf8\") " pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.844916 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17a6501d-7484-4265-9d97-f6c0b4582cf8-utilities\") pod \"redhat-operators-7q7pn\" (UID: \"17a6501d-7484-4265-9d97-f6c0b4582cf8\") " pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.844974 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17a6501d-7484-4265-9d97-f6c0b4582cf8-catalog-content\") pod \"redhat-operators-7q7pn\" (UID: \"17a6501d-7484-4265-9d97-f6c0b4582cf8\") " pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.845461 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17a6501d-7484-4265-9d97-f6c0b4582cf8-utilities\") pod \"redhat-operators-7q7pn\" (UID: \"17a6501d-7484-4265-9d97-f6c0b4582cf8\") " pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.845524 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17a6501d-7484-4265-9d97-f6c0b4582cf8-catalog-content\") pod \"redhat-operators-7q7pn\" (UID: \"17a6501d-7484-4265-9d97-f6c0b4582cf8\") " pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.878039 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgnvp\" (UniqueName: \"kubernetes.io/projected/17a6501d-7484-4265-9d97-f6c0b4582cf8-kube-api-access-pgnvp\") pod \"redhat-operators-7q7pn\" (UID: \"17a6501d-7484-4265-9d97-f6c0b4582cf8\") " pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:29 crc kubenswrapper[4808]: I1124 17:50:29.926633 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:30 crc kubenswrapper[4808]: I1124 17:50:30.398979 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7q7pn"] Nov 24 17:50:31 crc kubenswrapper[4808]: I1124 17:50:31.164445 4808 generic.go:334] "Generic (PLEG): container finished" podID="17a6501d-7484-4265-9d97-f6c0b4582cf8" containerID="b5e52b53587fa65a2b1fe7a0291ab4b866d0c941f4374e9558915eefb47938f8" exitCode=0 Nov 24 17:50:31 crc kubenswrapper[4808]: I1124 17:50:31.164551 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7q7pn" event={"ID":"17a6501d-7484-4265-9d97-f6c0b4582cf8","Type":"ContainerDied","Data":"b5e52b53587fa65a2b1fe7a0291ab4b866d0c941f4374e9558915eefb47938f8"} Nov 24 17:50:31 crc kubenswrapper[4808]: I1124 17:50:31.167920 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7q7pn" event={"ID":"17a6501d-7484-4265-9d97-f6c0b4582cf8","Type":"ContainerStarted","Data":"c5aec8251966f0f0074f229006b9616d9c40f34503eaa7ee4c26ccae45f01d10"} Nov 24 17:50:32 crc kubenswrapper[4808]: I1124 17:50:32.179689 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7q7pn" event={"ID":"17a6501d-7484-4265-9d97-f6c0b4582cf8","Type":"ContainerStarted","Data":"503290a7af6d08d6cea55d04a2453a4908f0cfa7a0f84bc32bdacf798c705755"} Nov 24 17:50:32 crc kubenswrapper[4808]: I1124 17:50:32.390456 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:32 crc kubenswrapper[4808]: I1124 17:50:32.390514 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:32 crc kubenswrapper[4808]: I1124 17:50:32.446152 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:33 crc kubenswrapper[4808]: I1124 17:50:33.232804 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:34 crc kubenswrapper[4808]: I1124 17:50:34.775339 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qn9nt"] Nov 24 17:50:35 crc kubenswrapper[4808]: I1124 17:50:35.213631 4808 generic.go:334] "Generic (PLEG): container finished" podID="17a6501d-7484-4265-9d97-f6c0b4582cf8" containerID="503290a7af6d08d6cea55d04a2453a4908f0cfa7a0f84bc32bdacf798c705755" exitCode=0 Nov 24 17:50:35 crc kubenswrapper[4808]: I1124 17:50:35.213720 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7q7pn" event={"ID":"17a6501d-7484-4265-9d97-f6c0b4582cf8","Type":"ContainerDied","Data":"503290a7af6d08d6cea55d04a2453a4908f0cfa7a0f84bc32bdacf798c705755"} Nov 24 17:50:35 crc kubenswrapper[4808]: I1124 17:50:35.214153 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qn9nt" podUID="ee37c403-358a-43db-912e-b04670d87178" containerName="registry-server" containerID="cri-o://da82936ec4da64c2709a3c1b6ca533dab8e307c442d6861f6ef37b52ec136e7f" gracePeriod=2 Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.226351 4808 generic.go:334] "Generic (PLEG): container finished" podID="ee37c403-358a-43db-912e-b04670d87178" containerID="da82936ec4da64c2709a3c1b6ca533dab8e307c442d6861f6ef37b52ec136e7f" exitCode=0 Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.226547 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qn9nt" event={"ID":"ee37c403-358a-43db-912e-b04670d87178","Type":"ContainerDied","Data":"da82936ec4da64c2709a3c1b6ca533dab8e307c442d6861f6ef37b52ec136e7f"} Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.522860 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.522923 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.522972 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.523740 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d1973d5d1c793927c6c85a086c241df6363c531bf6676a4e3fab54356b709ffb"} pod="openshift-machine-config-operator/machine-config-daemon-ht87w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.523816 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" containerID="cri-o://d1973d5d1c793927c6c85a086c241df6363c531bf6676a4e3fab54356b709ffb" gracePeriod=600 Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.714410 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.783978 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee37c403-358a-43db-912e-b04670d87178-catalog-content\") pod \"ee37c403-358a-43db-912e-b04670d87178\" (UID: \"ee37c403-358a-43db-912e-b04670d87178\") " Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.784089 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjhwf\" (UniqueName: \"kubernetes.io/projected/ee37c403-358a-43db-912e-b04670d87178-kube-api-access-jjhwf\") pod \"ee37c403-358a-43db-912e-b04670d87178\" (UID: \"ee37c403-358a-43db-912e-b04670d87178\") " Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.784146 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee37c403-358a-43db-912e-b04670d87178-utilities\") pod \"ee37c403-358a-43db-912e-b04670d87178\" (UID: \"ee37c403-358a-43db-912e-b04670d87178\") " Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.785005 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee37c403-358a-43db-912e-b04670d87178-utilities" (OuterVolumeSpecName: "utilities") pod "ee37c403-358a-43db-912e-b04670d87178" (UID: "ee37c403-358a-43db-912e-b04670d87178"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.791152 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee37c403-358a-43db-912e-b04670d87178-kube-api-access-jjhwf" (OuterVolumeSpecName: "kube-api-access-jjhwf") pod "ee37c403-358a-43db-912e-b04670d87178" (UID: "ee37c403-358a-43db-912e-b04670d87178"). InnerVolumeSpecName "kube-api-access-jjhwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.836792 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee37c403-358a-43db-912e-b04670d87178-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee37c403-358a-43db-912e-b04670d87178" (UID: "ee37c403-358a-43db-912e-b04670d87178"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.886572 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee37c403-358a-43db-912e-b04670d87178-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.886890 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjhwf\" (UniqueName: \"kubernetes.io/projected/ee37c403-358a-43db-912e-b04670d87178-kube-api-access-jjhwf\") on node \"crc\" DevicePath \"\"" Nov 24 17:50:36 crc kubenswrapper[4808]: I1124 17:50:36.886911 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee37c403-358a-43db-912e-b04670d87178-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:50:37 crc kubenswrapper[4808]: I1124 17:50:37.241547 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerID="d1973d5d1c793927c6c85a086c241df6363c531bf6676a4e3fab54356b709ffb" exitCode=0 Nov 24 17:50:37 crc kubenswrapper[4808]: I1124 17:50:37.241650 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerDied","Data":"d1973d5d1c793927c6c85a086c241df6363c531bf6676a4e3fab54356b709ffb"} Nov 24 17:50:37 crc kubenswrapper[4808]: I1124 17:50:37.241705 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f"} Nov 24 17:50:37 crc kubenswrapper[4808]: I1124 17:50:37.241723 4808 scope.go:117] "RemoveContainer" containerID="38d3a2110cebc7bb322ebd71a0ddbf64043c7ffce95b69f693d9dde312b5d138" Nov 24 17:50:37 crc kubenswrapper[4808]: I1124 17:50:37.243992 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7q7pn" event={"ID":"17a6501d-7484-4265-9d97-f6c0b4582cf8","Type":"ContainerStarted","Data":"79c105b79dd31bb3a3f5a4c7a8f4aea425c6eb2e03c85d494eebcc2bbfa5cad6"} Nov 24 17:50:37 crc kubenswrapper[4808]: I1124 17:50:37.246329 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qn9nt" event={"ID":"ee37c403-358a-43db-912e-b04670d87178","Type":"ContainerDied","Data":"d201dad0935d1fb58609c6a1fade8f9b79246d0433b9a920027110caa20bb7e1"} Nov 24 17:50:37 crc kubenswrapper[4808]: I1124 17:50:37.246397 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qn9nt" Nov 24 17:50:37 crc kubenswrapper[4808]: I1124 17:50:37.275380 4808 scope.go:117] "RemoveContainer" containerID="da82936ec4da64c2709a3c1b6ca533dab8e307c442d6861f6ef37b52ec136e7f" Nov 24 17:50:37 crc kubenswrapper[4808]: I1124 17:50:37.302221 4808 scope.go:117] "RemoveContainer" containerID="56074b5ea32d75ee4c355218d4303ea163ed1b2b286ed9f1b1c80df11ace495a" Nov 24 17:50:37 crc kubenswrapper[4808]: I1124 17:50:37.304289 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7q7pn" podStartSLOduration=3.431652343 podStartE2EDuration="8.304270641s" podCreationTimestamp="2025-11-24 17:50:29 +0000 UTC" firstStartedPulling="2025-11-24 17:50:31.167796299 +0000 UTC m=+1423.765464101" lastFinishedPulling="2025-11-24 17:50:36.040414597 +0000 UTC m=+1428.638082399" observedRunningTime="2025-11-24 17:50:37.283763332 +0000 UTC m=+1429.881431144" watchObservedRunningTime="2025-11-24 17:50:37.304270641 +0000 UTC m=+1429.901938443" Nov 24 17:50:37 crc kubenswrapper[4808]: I1124 17:50:37.305593 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qn9nt"] Nov 24 17:50:37 crc kubenswrapper[4808]: I1124 17:50:37.314183 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qn9nt"] Nov 24 17:50:37 crc kubenswrapper[4808]: I1124 17:50:37.334292 4808 scope.go:117] "RemoveContainer" containerID="63fb509837d612a84c7c247347a8c385822ba2ce8bef63384f0a733488c6a4d4" Nov 24 17:50:38 crc kubenswrapper[4808]: I1124 17:50:38.363245 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee37c403-358a-43db-912e-b04670d87178" path="/var/lib/kubelet/pods/ee37c403-358a-43db-912e-b04670d87178/volumes" Nov 24 17:50:39 crc kubenswrapper[4808]: I1124 17:50:39.927059 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:39 crc kubenswrapper[4808]: I1124 17:50:39.927416 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:40 crc kubenswrapper[4808]: I1124 17:50:40.977965 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7q7pn" podUID="17a6501d-7484-4265-9d97-f6c0b4582cf8" containerName="registry-server" probeResult="failure" output=< Nov 24 17:50:40 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Nov 24 17:50:40 crc kubenswrapper[4808]: > Nov 24 17:50:49 crc kubenswrapper[4808]: I1124 17:50:49.981673 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:50 crc kubenswrapper[4808]: I1124 17:50:50.033190 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:50 crc kubenswrapper[4808]: I1124 17:50:50.222666 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7q7pn"] Nov 24 17:50:51 crc kubenswrapper[4808]: I1124 17:50:51.387803 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7q7pn" podUID="17a6501d-7484-4265-9d97-f6c0b4582cf8" containerName="registry-server" containerID="cri-o://79c105b79dd31bb3a3f5a4c7a8f4aea425c6eb2e03c85d494eebcc2bbfa5cad6" gracePeriod=2 Nov 24 17:50:51 crc kubenswrapper[4808]: I1124 17:50:51.848703 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:51 crc kubenswrapper[4808]: I1124 17:50:51.981455 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17a6501d-7484-4265-9d97-f6c0b4582cf8-catalog-content\") pod \"17a6501d-7484-4265-9d97-f6c0b4582cf8\" (UID: \"17a6501d-7484-4265-9d97-f6c0b4582cf8\") " Nov 24 17:50:51 crc kubenswrapper[4808]: I1124 17:50:51.981530 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgnvp\" (UniqueName: \"kubernetes.io/projected/17a6501d-7484-4265-9d97-f6c0b4582cf8-kube-api-access-pgnvp\") pod \"17a6501d-7484-4265-9d97-f6c0b4582cf8\" (UID: \"17a6501d-7484-4265-9d97-f6c0b4582cf8\") " Nov 24 17:50:51 crc kubenswrapper[4808]: I1124 17:50:51.981585 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17a6501d-7484-4265-9d97-f6c0b4582cf8-utilities\") pod \"17a6501d-7484-4265-9d97-f6c0b4582cf8\" (UID: \"17a6501d-7484-4265-9d97-f6c0b4582cf8\") " Nov 24 17:50:51 crc kubenswrapper[4808]: I1124 17:50:51.982616 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17a6501d-7484-4265-9d97-f6c0b4582cf8-utilities" (OuterVolumeSpecName: "utilities") pod "17a6501d-7484-4265-9d97-f6c0b4582cf8" (UID: "17a6501d-7484-4265-9d97-f6c0b4582cf8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:50:51 crc kubenswrapper[4808]: I1124 17:50:51.988245 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17a6501d-7484-4265-9d97-f6c0b4582cf8-kube-api-access-pgnvp" (OuterVolumeSpecName: "kube-api-access-pgnvp") pod "17a6501d-7484-4265-9d97-f6c0b4582cf8" (UID: "17a6501d-7484-4265-9d97-f6c0b4582cf8"). InnerVolumeSpecName "kube-api-access-pgnvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.070534 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17a6501d-7484-4265-9d97-f6c0b4582cf8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17a6501d-7484-4265-9d97-f6c0b4582cf8" (UID: "17a6501d-7484-4265-9d97-f6c0b4582cf8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.083916 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17a6501d-7484-4265-9d97-f6c0b4582cf8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.083970 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgnvp\" (UniqueName: \"kubernetes.io/projected/17a6501d-7484-4265-9d97-f6c0b4582cf8-kube-api-access-pgnvp\") on node \"crc\" DevicePath \"\"" Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.083986 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17a6501d-7484-4265-9d97-f6c0b4582cf8-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.398747 4808 generic.go:334] "Generic (PLEG): container finished" podID="17a6501d-7484-4265-9d97-f6c0b4582cf8" containerID="79c105b79dd31bb3a3f5a4c7a8f4aea425c6eb2e03c85d494eebcc2bbfa5cad6" exitCode=0 Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.398984 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7q7pn" event={"ID":"17a6501d-7484-4265-9d97-f6c0b4582cf8","Type":"ContainerDied","Data":"79c105b79dd31bb3a3f5a4c7a8f4aea425c6eb2e03c85d494eebcc2bbfa5cad6"} Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.399110 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7q7pn" Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.399138 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7q7pn" event={"ID":"17a6501d-7484-4265-9d97-f6c0b4582cf8","Type":"ContainerDied","Data":"c5aec8251966f0f0074f229006b9616d9c40f34503eaa7ee4c26ccae45f01d10"} Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.399171 4808 scope.go:117] "RemoveContainer" containerID="79c105b79dd31bb3a3f5a4c7a8f4aea425c6eb2e03c85d494eebcc2bbfa5cad6" Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.423239 4808 scope.go:117] "RemoveContainer" containerID="503290a7af6d08d6cea55d04a2453a4908f0cfa7a0f84bc32bdacf798c705755" Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.435196 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7q7pn"] Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.442761 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7q7pn"] Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.457576 4808 scope.go:117] "RemoveContainer" containerID="b5e52b53587fa65a2b1fe7a0291ab4b866d0c941f4374e9558915eefb47938f8" Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.501919 4808 scope.go:117] "RemoveContainer" containerID="79c105b79dd31bb3a3f5a4c7a8f4aea425c6eb2e03c85d494eebcc2bbfa5cad6" Nov 24 17:50:52 crc kubenswrapper[4808]: E1124 17:50:52.502350 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79c105b79dd31bb3a3f5a4c7a8f4aea425c6eb2e03c85d494eebcc2bbfa5cad6\": container with ID starting with 79c105b79dd31bb3a3f5a4c7a8f4aea425c6eb2e03c85d494eebcc2bbfa5cad6 not found: ID does not exist" containerID="79c105b79dd31bb3a3f5a4c7a8f4aea425c6eb2e03c85d494eebcc2bbfa5cad6" Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.502389 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79c105b79dd31bb3a3f5a4c7a8f4aea425c6eb2e03c85d494eebcc2bbfa5cad6"} err="failed to get container status \"79c105b79dd31bb3a3f5a4c7a8f4aea425c6eb2e03c85d494eebcc2bbfa5cad6\": rpc error: code = NotFound desc = could not find container \"79c105b79dd31bb3a3f5a4c7a8f4aea425c6eb2e03c85d494eebcc2bbfa5cad6\": container with ID starting with 79c105b79dd31bb3a3f5a4c7a8f4aea425c6eb2e03c85d494eebcc2bbfa5cad6 not found: ID does not exist" Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.502415 4808 scope.go:117] "RemoveContainer" containerID="503290a7af6d08d6cea55d04a2453a4908f0cfa7a0f84bc32bdacf798c705755" Nov 24 17:50:52 crc kubenswrapper[4808]: E1124 17:50:52.502888 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"503290a7af6d08d6cea55d04a2453a4908f0cfa7a0f84bc32bdacf798c705755\": container with ID starting with 503290a7af6d08d6cea55d04a2453a4908f0cfa7a0f84bc32bdacf798c705755 not found: ID does not exist" containerID="503290a7af6d08d6cea55d04a2453a4908f0cfa7a0f84bc32bdacf798c705755" Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.502913 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"503290a7af6d08d6cea55d04a2453a4908f0cfa7a0f84bc32bdacf798c705755"} err="failed to get container status \"503290a7af6d08d6cea55d04a2453a4908f0cfa7a0f84bc32bdacf798c705755\": rpc error: code = NotFound desc = could not find container \"503290a7af6d08d6cea55d04a2453a4908f0cfa7a0f84bc32bdacf798c705755\": container with ID starting with 503290a7af6d08d6cea55d04a2453a4908f0cfa7a0f84bc32bdacf798c705755 not found: ID does not exist" Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.502931 4808 scope.go:117] "RemoveContainer" containerID="b5e52b53587fa65a2b1fe7a0291ab4b866d0c941f4374e9558915eefb47938f8" Nov 24 17:50:52 crc kubenswrapper[4808]: E1124 17:50:52.503272 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5e52b53587fa65a2b1fe7a0291ab4b866d0c941f4374e9558915eefb47938f8\": container with ID starting with b5e52b53587fa65a2b1fe7a0291ab4b866d0c941f4374e9558915eefb47938f8 not found: ID does not exist" containerID="b5e52b53587fa65a2b1fe7a0291ab4b866d0c941f4374e9558915eefb47938f8" Nov 24 17:50:52 crc kubenswrapper[4808]: I1124 17:50:52.503315 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5e52b53587fa65a2b1fe7a0291ab4b866d0c941f4374e9558915eefb47938f8"} err="failed to get container status \"b5e52b53587fa65a2b1fe7a0291ab4b866d0c941f4374e9558915eefb47938f8\": rpc error: code = NotFound desc = could not find container \"b5e52b53587fa65a2b1fe7a0291ab4b866d0c941f4374e9558915eefb47938f8\": container with ID starting with b5e52b53587fa65a2b1fe7a0291ab4b866d0c941f4374e9558915eefb47938f8 not found: ID does not exist" Nov 24 17:50:54 crc kubenswrapper[4808]: I1124 17:50:54.358365 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17a6501d-7484-4265-9d97-f6c0b4582cf8" path="/var/lib/kubelet/pods/17a6501d-7484-4265-9d97-f6c0b4582cf8/volumes" Nov 24 17:51:02 crc kubenswrapper[4808]: I1124 17:51:02.492326 4808 generic.go:334] "Generic (PLEG): container finished" podID="b6ec6116-79f6-4178-9ebc-bc29c07889b3" containerID="4a2b93b0c0722322a99cc63b1c97a9fac25ca1f5e13ff64e848a3d7e58830e85" exitCode=0 Nov 24 17:51:02 crc kubenswrapper[4808]: I1124 17:51:02.492409 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" event={"ID":"b6ec6116-79f6-4178-9ebc-bc29c07889b3","Type":"ContainerDied","Data":"4a2b93b0c0722322a99cc63b1c97a9fac25ca1f5e13ff64e848a3d7e58830e85"} Nov 24 17:51:03 crc kubenswrapper[4808]: I1124 17:51:03.900427 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.042559 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-ssh-key\") pod \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.042727 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-inventory\") pod \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.042760 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-bootstrap-combined-ca-bundle\") pod \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.042827 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkkdq\" (UniqueName: \"kubernetes.io/projected/b6ec6116-79f6-4178-9ebc-bc29c07889b3-kube-api-access-nkkdq\") pod \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\" (UID: \"b6ec6116-79f6-4178-9ebc-bc29c07889b3\") " Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.048331 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b6ec6116-79f6-4178-9ebc-bc29c07889b3" (UID: "b6ec6116-79f6-4178-9ebc-bc29c07889b3"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.048423 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6ec6116-79f6-4178-9ebc-bc29c07889b3-kube-api-access-nkkdq" (OuterVolumeSpecName: "kube-api-access-nkkdq") pod "b6ec6116-79f6-4178-9ebc-bc29c07889b3" (UID: "b6ec6116-79f6-4178-9ebc-bc29c07889b3"). InnerVolumeSpecName "kube-api-access-nkkdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.069554 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b6ec6116-79f6-4178-9ebc-bc29c07889b3" (UID: "b6ec6116-79f6-4178-9ebc-bc29c07889b3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.074775 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-inventory" (OuterVolumeSpecName: "inventory") pod "b6ec6116-79f6-4178-9ebc-bc29c07889b3" (UID: "b6ec6116-79f6-4178-9ebc-bc29c07889b3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.164154 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.164189 4808 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.164202 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkkdq\" (UniqueName: \"kubernetes.io/projected/b6ec6116-79f6-4178-9ebc-bc29c07889b3-kube-api-access-nkkdq\") on node \"crc\" DevicePath \"\"" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.164212 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6ec6116-79f6-4178-9ebc-bc29c07889b3-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.510186 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" event={"ID":"b6ec6116-79f6-4178-9ebc-bc29c07889b3","Type":"ContainerDied","Data":"e36e94e3708a372fc24d1ce415cb58793fd1823e8d08d0e569236cdd54ed31cc"} Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.510236 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e36e94e3708a372fc24d1ce415cb58793fd1823e8d08d0e569236cdd54ed31cc" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.510269 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.602905 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85"] Nov 24 17:51:04 crc kubenswrapper[4808]: E1124 17:51:04.603362 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee37c403-358a-43db-912e-b04670d87178" containerName="registry-server" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.603378 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee37c403-358a-43db-912e-b04670d87178" containerName="registry-server" Nov 24 17:51:04 crc kubenswrapper[4808]: E1124 17:51:04.603393 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee37c403-358a-43db-912e-b04670d87178" containerName="extract-utilities" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.603399 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee37c403-358a-43db-912e-b04670d87178" containerName="extract-utilities" Nov 24 17:51:04 crc kubenswrapper[4808]: E1124 17:51:04.603413 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17a6501d-7484-4265-9d97-f6c0b4582cf8" containerName="extract-utilities" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.603420 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="17a6501d-7484-4265-9d97-f6c0b4582cf8" containerName="extract-utilities" Nov 24 17:51:04 crc kubenswrapper[4808]: E1124 17:51:04.603433 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee37c403-358a-43db-912e-b04670d87178" containerName="extract-content" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.603440 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee37c403-358a-43db-912e-b04670d87178" containerName="extract-content" Nov 24 17:51:04 crc kubenswrapper[4808]: E1124 17:51:04.603449 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6ec6116-79f6-4178-9ebc-bc29c07889b3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.603455 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6ec6116-79f6-4178-9ebc-bc29c07889b3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 17:51:04 crc kubenswrapper[4808]: E1124 17:51:04.603480 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17a6501d-7484-4265-9d97-f6c0b4582cf8" containerName="extract-content" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.603486 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="17a6501d-7484-4265-9d97-f6c0b4582cf8" containerName="extract-content" Nov 24 17:51:04 crc kubenswrapper[4808]: E1124 17:51:04.603499 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17a6501d-7484-4265-9d97-f6c0b4582cf8" containerName="registry-server" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.603505 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="17a6501d-7484-4265-9d97-f6c0b4582cf8" containerName="registry-server" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.603703 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6ec6116-79f6-4178-9ebc-bc29c07889b3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.603714 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="17a6501d-7484-4265-9d97-f6c0b4582cf8" containerName="registry-server" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.603724 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee37c403-358a-43db-912e-b04670d87178" containerName="registry-server" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.604414 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.606644 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.606735 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.608044 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.610387 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.618767 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85"] Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.675371 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee20a194-4626-4fe3-8138-e103a44a2006-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s4f85\" (UID: \"ee20a194-4626-4fe3-8138-e103a44a2006\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.675490 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7sjm\" (UniqueName: \"kubernetes.io/projected/ee20a194-4626-4fe3-8138-e103a44a2006-kube-api-access-v7sjm\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s4f85\" (UID: \"ee20a194-4626-4fe3-8138-e103a44a2006\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.675678 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee20a194-4626-4fe3-8138-e103a44a2006-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s4f85\" (UID: \"ee20a194-4626-4fe3-8138-e103a44a2006\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.777608 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7sjm\" (UniqueName: \"kubernetes.io/projected/ee20a194-4626-4fe3-8138-e103a44a2006-kube-api-access-v7sjm\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s4f85\" (UID: \"ee20a194-4626-4fe3-8138-e103a44a2006\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.778003 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee20a194-4626-4fe3-8138-e103a44a2006-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s4f85\" (UID: \"ee20a194-4626-4fe3-8138-e103a44a2006\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.778142 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee20a194-4626-4fe3-8138-e103a44a2006-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s4f85\" (UID: \"ee20a194-4626-4fe3-8138-e103a44a2006\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.782606 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee20a194-4626-4fe3-8138-e103a44a2006-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s4f85\" (UID: \"ee20a194-4626-4fe3-8138-e103a44a2006\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.790726 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee20a194-4626-4fe3-8138-e103a44a2006-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s4f85\" (UID: \"ee20a194-4626-4fe3-8138-e103a44a2006\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.793796 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7sjm\" (UniqueName: \"kubernetes.io/projected/ee20a194-4626-4fe3-8138-e103a44a2006-kube-api-access-v7sjm\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s4f85\" (UID: \"ee20a194-4626-4fe3-8138-e103a44a2006\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" Nov 24 17:51:04 crc kubenswrapper[4808]: I1124 17:51:04.922668 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" Nov 24 17:51:05 crc kubenswrapper[4808]: I1124 17:51:05.482791 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85"] Nov 24 17:51:05 crc kubenswrapper[4808]: W1124 17:51:05.485649 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee20a194_4626_4fe3_8138_e103a44a2006.slice/crio-ccdd2c5a889f5baff4de2273cea225d16232ce77ac1b836e8399131904952ac0 WatchSource:0}: Error finding container ccdd2c5a889f5baff4de2273cea225d16232ce77ac1b836e8399131904952ac0: Status 404 returned error can't find the container with id ccdd2c5a889f5baff4de2273cea225d16232ce77ac1b836e8399131904952ac0 Nov 24 17:51:05 crc kubenswrapper[4808]: I1124 17:51:05.519735 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" event={"ID":"ee20a194-4626-4fe3-8138-e103a44a2006","Type":"ContainerStarted","Data":"ccdd2c5a889f5baff4de2273cea225d16232ce77ac1b836e8399131904952ac0"} Nov 24 17:51:06 crc kubenswrapper[4808]: I1124 17:51:06.530608 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" event={"ID":"ee20a194-4626-4fe3-8138-e103a44a2006","Type":"ContainerStarted","Data":"3024de62b5e63a22e95eda2d57a5cc9b7dbce52a11ff47acd7602d969bc1ce3b"} Nov 24 17:51:06 crc kubenswrapper[4808]: I1124 17:51:06.552820 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" podStartSLOduration=2.05177772 podStartE2EDuration="2.552800132s" podCreationTimestamp="2025-11-24 17:51:04 +0000 UTC" firstStartedPulling="2025-11-24 17:51:05.48764222 +0000 UTC m=+1458.085310022" lastFinishedPulling="2025-11-24 17:51:05.988664632 +0000 UTC m=+1458.586332434" observedRunningTime="2025-11-24 17:51:06.54551373 +0000 UTC m=+1459.143181532" watchObservedRunningTime="2025-11-24 17:51:06.552800132 +0000 UTC m=+1459.150467924" Nov 24 17:51:52 crc kubenswrapper[4808]: I1124 17:51:52.500386 4808 scope.go:117] "RemoveContainer" containerID="07ce607292a080bb811b46df930b8b0039a46023a74a45bff26944dc5fa5c773" Nov 24 17:51:52 crc kubenswrapper[4808]: I1124 17:51:52.525325 4808 scope.go:117] "RemoveContainer" containerID="84f7d4e24d959ce8f11afb76b8220e2fb780828e8701136618a2be208a1d5dfa" Nov 24 17:52:14 crc kubenswrapper[4808]: I1124 17:52:14.036502 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-8z24k"] Nov 24 17:52:14 crc kubenswrapper[4808]: I1124 17:52:14.045470 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-b750-account-create-update-v7jcb"] Nov 24 17:52:14 crc kubenswrapper[4808]: I1124 17:52:14.056047 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-8z24k"] Nov 24 17:52:14 crc kubenswrapper[4808]: I1124 17:52:14.064435 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-b750-account-create-update-v7jcb"] Nov 24 17:52:14 crc kubenswrapper[4808]: I1124 17:52:14.360143 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3563a459-e344-43c7-83ee-3da790cfe660" path="/var/lib/kubelet/pods/3563a459-e344-43c7-83ee-3da790cfe660/volumes" Nov 24 17:52:14 crc kubenswrapper[4808]: I1124 17:52:14.361576 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c" path="/var/lib/kubelet/pods/86f2ee9f-f7b0-46bc-8a7a-f8af03f3989c/volumes" Nov 24 17:52:18 crc kubenswrapper[4808]: I1124 17:52:18.033055 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5213-account-create-update-q7rbp"] Nov 24 17:52:18 crc kubenswrapper[4808]: I1124 17:52:18.049237 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-fmsjw"] Nov 24 17:52:18 crc kubenswrapper[4808]: I1124 17:52:18.059528 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6ab9-account-create-update-x9n7t"] Nov 24 17:52:18 crc kubenswrapper[4808]: I1124 17:52:18.067637 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5213-account-create-update-q7rbp"] Nov 24 17:52:18 crc kubenswrapper[4808]: I1124 17:52:18.076111 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-fmsjw"] Nov 24 17:52:18 crc kubenswrapper[4808]: I1124 17:52:18.083488 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-6ab9-account-create-update-x9n7t"] Nov 24 17:52:18 crc kubenswrapper[4808]: I1124 17:52:18.090984 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-9jp7n"] Nov 24 17:52:18 crc kubenswrapper[4808]: I1124 17:52:18.098659 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-9jp7n"] Nov 24 17:52:18 crc kubenswrapper[4808]: I1124 17:52:18.360000 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39438be0-89df-4dff-98f3-368cb216fb6c" path="/var/lib/kubelet/pods/39438be0-89df-4dff-98f3-368cb216fb6c/volumes" Nov 24 17:52:18 crc kubenswrapper[4808]: I1124 17:52:18.360683 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0" path="/var/lib/kubelet/pods/6ea9f9cd-1cdc-481a-b878-4d7892e9a8d0/volumes" Nov 24 17:52:18 crc kubenswrapper[4808]: I1124 17:52:18.361380 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c2587b0-bdda-4877-8847-c8af2afb7735" path="/var/lib/kubelet/pods/7c2587b0-bdda-4877-8847-c8af2afb7735/volumes" Nov 24 17:52:18 crc kubenswrapper[4808]: I1124 17:52:18.361985 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef8985b9-2261-40cf-bfa9-ef86fb447616" path="/var/lib/kubelet/pods/ef8985b9-2261-40cf-bfa9-ef86fb447616/volumes" Nov 24 17:52:33 crc kubenswrapper[4808]: I1124 17:52:33.305720 4808 generic.go:334] "Generic (PLEG): container finished" podID="ee20a194-4626-4fe3-8138-e103a44a2006" containerID="3024de62b5e63a22e95eda2d57a5cc9b7dbce52a11ff47acd7602d969bc1ce3b" exitCode=0 Nov 24 17:52:33 crc kubenswrapper[4808]: I1124 17:52:33.305821 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" event={"ID":"ee20a194-4626-4fe3-8138-e103a44a2006","Type":"ContainerDied","Data":"3024de62b5e63a22e95eda2d57a5cc9b7dbce52a11ff47acd7602d969bc1ce3b"} Nov 24 17:52:34 crc kubenswrapper[4808]: I1124 17:52:34.737673 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" Nov 24 17:52:34 crc kubenswrapper[4808]: I1124 17:52:34.846405 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7sjm\" (UniqueName: \"kubernetes.io/projected/ee20a194-4626-4fe3-8138-e103a44a2006-kube-api-access-v7sjm\") pod \"ee20a194-4626-4fe3-8138-e103a44a2006\" (UID: \"ee20a194-4626-4fe3-8138-e103a44a2006\") " Nov 24 17:52:34 crc kubenswrapper[4808]: I1124 17:52:34.846503 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee20a194-4626-4fe3-8138-e103a44a2006-inventory\") pod \"ee20a194-4626-4fe3-8138-e103a44a2006\" (UID: \"ee20a194-4626-4fe3-8138-e103a44a2006\") " Nov 24 17:52:34 crc kubenswrapper[4808]: I1124 17:52:34.846637 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee20a194-4626-4fe3-8138-e103a44a2006-ssh-key\") pod \"ee20a194-4626-4fe3-8138-e103a44a2006\" (UID: \"ee20a194-4626-4fe3-8138-e103a44a2006\") " Nov 24 17:52:34 crc kubenswrapper[4808]: I1124 17:52:34.854299 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee20a194-4626-4fe3-8138-e103a44a2006-kube-api-access-v7sjm" (OuterVolumeSpecName: "kube-api-access-v7sjm") pod "ee20a194-4626-4fe3-8138-e103a44a2006" (UID: "ee20a194-4626-4fe3-8138-e103a44a2006"). InnerVolumeSpecName "kube-api-access-v7sjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:52:34 crc kubenswrapper[4808]: I1124 17:52:34.877071 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee20a194-4626-4fe3-8138-e103a44a2006-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ee20a194-4626-4fe3-8138-e103a44a2006" (UID: "ee20a194-4626-4fe3-8138-e103a44a2006"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:52:34 crc kubenswrapper[4808]: I1124 17:52:34.877653 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee20a194-4626-4fe3-8138-e103a44a2006-inventory" (OuterVolumeSpecName: "inventory") pod "ee20a194-4626-4fe3-8138-e103a44a2006" (UID: "ee20a194-4626-4fe3-8138-e103a44a2006"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:52:34 crc kubenswrapper[4808]: I1124 17:52:34.966389 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7sjm\" (UniqueName: \"kubernetes.io/projected/ee20a194-4626-4fe3-8138-e103a44a2006-kube-api-access-v7sjm\") on node \"crc\" DevicePath \"\"" Nov 24 17:52:34 crc kubenswrapper[4808]: I1124 17:52:34.966448 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee20a194-4626-4fe3-8138-e103a44a2006-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:52:34 crc kubenswrapper[4808]: I1124 17:52:34.966460 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee20a194-4626-4fe3-8138-e103a44a2006-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.325623 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" event={"ID":"ee20a194-4626-4fe3-8138-e103a44a2006","Type":"ContainerDied","Data":"ccdd2c5a889f5baff4de2273cea225d16232ce77ac1b836e8399131904952ac0"} Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.325668 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccdd2c5a889f5baff4de2273cea225d16232ce77ac1b836e8399131904952ac0" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.325741 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s4f85" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.411283 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk"] Nov 24 17:52:35 crc kubenswrapper[4808]: E1124 17:52:35.411759 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee20a194-4626-4fe3-8138-e103a44a2006" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.411779 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee20a194-4626-4fe3-8138-e103a44a2006" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.411977 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee20a194-4626-4fe3-8138-e103a44a2006" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.412783 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.416653 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.416852 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.417210 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.417515 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.425840 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk"] Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.578680 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91323238-4f1e-4dd3-84f2-6fc8bf252642-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wglmk\" (UID: \"91323238-4f1e-4dd3-84f2-6fc8bf252642\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.579223 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91323238-4f1e-4dd3-84f2-6fc8bf252642-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wglmk\" (UID: \"91323238-4f1e-4dd3-84f2-6fc8bf252642\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.579282 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w62dt\" (UniqueName: \"kubernetes.io/projected/91323238-4f1e-4dd3-84f2-6fc8bf252642-kube-api-access-w62dt\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wglmk\" (UID: \"91323238-4f1e-4dd3-84f2-6fc8bf252642\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.680934 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91323238-4f1e-4dd3-84f2-6fc8bf252642-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wglmk\" (UID: \"91323238-4f1e-4dd3-84f2-6fc8bf252642\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.681002 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91323238-4f1e-4dd3-84f2-6fc8bf252642-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wglmk\" (UID: \"91323238-4f1e-4dd3-84f2-6fc8bf252642\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.681064 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w62dt\" (UniqueName: \"kubernetes.io/projected/91323238-4f1e-4dd3-84f2-6fc8bf252642-kube-api-access-w62dt\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wglmk\" (UID: \"91323238-4f1e-4dd3-84f2-6fc8bf252642\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.687509 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91323238-4f1e-4dd3-84f2-6fc8bf252642-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wglmk\" (UID: \"91323238-4f1e-4dd3-84f2-6fc8bf252642\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.688070 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91323238-4f1e-4dd3-84f2-6fc8bf252642-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wglmk\" (UID: \"91323238-4f1e-4dd3-84f2-6fc8bf252642\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.699780 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w62dt\" (UniqueName: \"kubernetes.io/projected/91323238-4f1e-4dd3-84f2-6fc8bf252642-kube-api-access-w62dt\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wglmk\" (UID: \"91323238-4f1e-4dd3-84f2-6fc8bf252642\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" Nov 24 17:52:35 crc kubenswrapper[4808]: I1124 17:52:35.733166 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" Nov 24 17:52:36 crc kubenswrapper[4808]: I1124 17:52:36.296696 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:52:36 crc kubenswrapper[4808]: I1124 17:52:36.303953 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk"] Nov 24 17:52:36 crc kubenswrapper[4808]: I1124 17:52:36.334139 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" event={"ID":"91323238-4f1e-4dd3-84f2-6fc8bf252642","Type":"ContainerStarted","Data":"a282b26fbd55afcdfc0b6336c684ec2ce64e0e334ac6c496d7728653b0a8b7c9"} Nov 24 17:52:36 crc kubenswrapper[4808]: I1124 17:52:36.522652 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:52:36 crc kubenswrapper[4808]: I1124 17:52:36.523104 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:52:38 crc kubenswrapper[4808]: I1124 17:52:38.358363 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" event={"ID":"91323238-4f1e-4dd3-84f2-6fc8bf252642","Type":"ContainerStarted","Data":"92f2b8968dfc4e86050d852528d3c6677c69b6304cd5db0d86459b38da584526"} Nov 24 17:52:38 crc kubenswrapper[4808]: I1124 17:52:38.419948 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" podStartSLOduration=2.517688556 podStartE2EDuration="3.419927584s" podCreationTimestamp="2025-11-24 17:52:35 +0000 UTC" firstStartedPulling="2025-11-24 17:52:36.296376895 +0000 UTC m=+1548.894044697" lastFinishedPulling="2025-11-24 17:52:37.198615923 +0000 UTC m=+1549.796283725" observedRunningTime="2025-11-24 17:52:38.412647002 +0000 UTC m=+1551.010314804" watchObservedRunningTime="2025-11-24 17:52:38.419927584 +0000 UTC m=+1551.017595386" Nov 24 17:52:42 crc kubenswrapper[4808]: I1124 17:52:42.039901 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-d6rlt"] Nov 24 17:52:42 crc kubenswrapper[4808]: I1124 17:52:42.049267 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-d6rlt"] Nov 24 17:52:42 crc kubenswrapper[4808]: I1124 17:52:42.364618 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="813f14a2-3982-4e6d-aca7-109d2bf2d1cb" path="/var/lib/kubelet/pods/813f14a2-3982-4e6d-aca7-109d2bf2d1cb/volumes" Nov 24 17:52:49 crc kubenswrapper[4808]: I1124 17:52:49.047144 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-dr57t"] Nov 24 17:52:49 crc kubenswrapper[4808]: I1124 17:52:49.062479 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-qg8hc"] Nov 24 17:52:49 crc kubenswrapper[4808]: I1124 17:52:49.074720 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-df60-account-create-update-mlxn7"] Nov 24 17:52:49 crc kubenswrapper[4808]: I1124 17:52:49.088323 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-b2sdj"] Nov 24 17:52:49 crc kubenswrapper[4808]: I1124 17:52:49.097205 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-cb38-account-create-update-zdknp"] Nov 24 17:52:49 crc kubenswrapper[4808]: I1124 17:52:49.106420 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-df60-account-create-update-mlxn7"] Nov 24 17:52:49 crc kubenswrapper[4808]: I1124 17:52:49.118393 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-dr57t"] Nov 24 17:52:49 crc kubenswrapper[4808]: I1124 17:52:49.128209 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-cb38-account-create-update-zdknp"] Nov 24 17:52:49 crc kubenswrapper[4808]: I1124 17:52:49.136935 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-qg8hc"] Nov 24 17:52:49 crc kubenswrapper[4808]: I1124 17:52:49.147356 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-b2sdj"] Nov 24 17:52:49 crc kubenswrapper[4808]: I1124 17:52:49.157709 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d41e-account-create-update-jm4rk"] Nov 24 17:52:49 crc kubenswrapper[4808]: I1124 17:52:49.167046 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-d41e-account-create-update-jm4rk"] Nov 24 17:52:50 crc kubenswrapper[4808]: I1124 17:52:50.366479 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01e65cec-01ec-4374-b9e7-c1b13f9b8d5e" path="/var/lib/kubelet/pods/01e65cec-01ec-4374-b9e7-c1b13f9b8d5e/volumes" Nov 24 17:52:50 crc kubenswrapper[4808]: I1124 17:52:50.367761 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04e46a77-a53c-4edd-a1d8-c74a7b61c666" path="/var/lib/kubelet/pods/04e46a77-a53c-4edd-a1d8-c74a7b61c666/volumes" Nov 24 17:52:50 crc kubenswrapper[4808]: I1124 17:52:50.368729 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="180dfd5f-4d34-42fc-983d-90a54e945a15" path="/var/lib/kubelet/pods/180dfd5f-4d34-42fc-983d-90a54e945a15/volumes" Nov 24 17:52:50 crc kubenswrapper[4808]: I1124 17:52:50.369629 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a" path="/var/lib/kubelet/pods/53e58aa9-1f2c-40e0-a608-e5f6c0e8d84a/volumes" Nov 24 17:52:50 crc kubenswrapper[4808]: I1124 17:52:50.370821 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5df6e93-265c-4ff6-8bf5-974e1bc7c742" path="/var/lib/kubelet/pods/c5df6e93-265c-4ff6-8bf5-974e1bc7c742/volumes" Nov 24 17:52:50 crc kubenswrapper[4808]: I1124 17:52:50.371511 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f113a2b1-1597-48d7-9420-16d6f34fedf8" path="/var/lib/kubelet/pods/f113a2b1-1597-48d7-9420-16d6f34fedf8/volumes" Nov 24 17:52:52 crc kubenswrapper[4808]: I1124 17:52:52.596203 4808 scope.go:117] "RemoveContainer" containerID="77f633b41f6fa0b20850e9facbfc52c1e7860f9a15833d8b38b86789c1e895e1" Nov 24 17:52:52 crc kubenswrapper[4808]: I1124 17:52:52.638082 4808 scope.go:117] "RemoveContainer" containerID="0581e349c664de7c6afa7e52742d3c29dea51433c46495c3e6012ba9da3c60e4" Nov 24 17:52:52 crc kubenswrapper[4808]: I1124 17:52:52.706996 4808 scope.go:117] "RemoveContainer" containerID="d975a76363a71b34015a1762ce112417a4bde8399ede3829500729785f099064" Nov 24 17:52:52 crc kubenswrapper[4808]: I1124 17:52:52.748598 4808 scope.go:117] "RemoveContainer" containerID="df8384c369f4dc156269d9d1914456e7b26d80de43e0054abef15709835f7c66" Nov 24 17:52:52 crc kubenswrapper[4808]: I1124 17:52:52.817823 4808 scope.go:117] "RemoveContainer" containerID="bda72b150f53176d32dabd034c30198f9b3df9da8544fb4ad46c3f00f23fc9b7" Nov 24 17:52:52 crc kubenswrapper[4808]: I1124 17:52:52.858394 4808 scope.go:117] "RemoveContainer" containerID="20fc52ba1f354e04df85fd49f44578b343421e2d26d9e773d0051d5d4f845b1e" Nov 24 17:52:52 crc kubenswrapper[4808]: I1124 17:52:52.898795 4808 scope.go:117] "RemoveContainer" containerID="52ed0d373b7440bd7b4bebe6c7bb72d50e7aa0b08704410979d013e188fa72d9" Nov 24 17:52:52 crc kubenswrapper[4808]: I1124 17:52:52.919607 4808 scope.go:117] "RemoveContainer" containerID="8df39cb91c81ad1cd4d30c723b62d14131e8133471312802206d869b1a6602d6" Nov 24 17:52:52 crc kubenswrapper[4808]: I1124 17:52:52.943913 4808 scope.go:117] "RemoveContainer" containerID="b44d74ea56f05888af192234ad258fc574a05e31916e6b4004f52c0fc779444b" Nov 24 17:52:52 crc kubenswrapper[4808]: I1124 17:52:52.963047 4808 scope.go:117] "RemoveContainer" containerID="646d8e94f3759ceb3ec9c2a8ebd53163ad4acb5b28a241abafec5e92f2230717" Nov 24 17:52:52 crc kubenswrapper[4808]: I1124 17:52:52.996069 4808 scope.go:117] "RemoveContainer" containerID="2bd53b4eed9898815b0999e1fd92e3a9df0481a425074def41bd8f58099737d5" Nov 24 17:52:53 crc kubenswrapper[4808]: I1124 17:52:53.029447 4808 scope.go:117] "RemoveContainer" containerID="7c668f7132a032ed325d0ad3fd827286c17794f129ae67da35df367aaa4aaf55" Nov 24 17:52:53 crc kubenswrapper[4808]: I1124 17:52:53.050598 4808 scope.go:117] "RemoveContainer" containerID="0479a82d98c3c1e45a143d1dff8827a025530872bb39af91ac7b4db3e76d6ded" Nov 24 17:52:54 crc kubenswrapper[4808]: I1124 17:52:54.030252 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-5bhqp"] Nov 24 17:52:54 crc kubenswrapper[4808]: I1124 17:52:54.041660 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-5bhqp"] Nov 24 17:52:54 crc kubenswrapper[4808]: I1124 17:52:54.357941 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ebabded-4179-4586-9be7-40d4635f7806" path="/var/lib/kubelet/pods/7ebabded-4179-4586-9be7-40d4635f7806/volumes" Nov 24 17:53:06 crc kubenswrapper[4808]: I1124 17:53:06.523276 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:53:06 crc kubenswrapper[4808]: I1124 17:53:06.524212 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:53:36 crc kubenswrapper[4808]: I1124 17:53:36.522776 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:53:36 crc kubenswrapper[4808]: I1124 17:53:36.523385 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:53:36 crc kubenswrapper[4808]: I1124 17:53:36.523446 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 17:53:36 crc kubenswrapper[4808]: I1124 17:53:36.524438 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f"} pod="openshift-machine-config-operator/machine-config-daemon-ht87w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:53:36 crc kubenswrapper[4808]: I1124 17:53:36.524530 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" containerID="cri-o://84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" gracePeriod=600 Nov 24 17:53:36 crc kubenswrapper[4808]: E1124 17:53:36.652383 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:53:36 crc kubenswrapper[4808]: I1124 17:53:36.904363 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" exitCode=0 Nov 24 17:53:36 crc kubenswrapper[4808]: I1124 17:53:36.904406 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerDied","Data":"84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f"} Nov 24 17:53:36 crc kubenswrapper[4808]: I1124 17:53:36.904443 4808 scope.go:117] "RemoveContainer" containerID="d1973d5d1c793927c6c85a086c241df6363c531bf6676a4e3fab54356b709ffb" Nov 24 17:53:36 crc kubenswrapper[4808]: I1124 17:53:36.904967 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:53:36 crc kubenswrapper[4808]: E1124 17:53:36.905284 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:53:40 crc kubenswrapper[4808]: I1124 17:53:40.046118 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dj9zb"] Nov 24 17:53:40 crc kubenswrapper[4808]: I1124 17:53:40.054120 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dj9zb"] Nov 24 17:53:40 crc kubenswrapper[4808]: I1124 17:53:40.359511 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56775e00-b320-4260-b162-e99c948df631" path="/var/lib/kubelet/pods/56775e00-b320-4260-b162-e99c948df631/volumes" Nov 24 17:53:41 crc kubenswrapper[4808]: I1124 17:53:41.030297 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-kp6v5"] Nov 24 17:53:41 crc kubenswrapper[4808]: I1124 17:53:41.039725 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-kp6v5"] Nov 24 17:53:42 crc kubenswrapper[4808]: I1124 17:53:42.368986 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2112bdd-0811-4bd0-927d-277b44b5cfc0" path="/var/lib/kubelet/pods/b2112bdd-0811-4bd0-927d-277b44b5cfc0/volumes" Nov 24 17:53:43 crc kubenswrapper[4808]: I1124 17:53:43.027870 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-wjrlj"] Nov 24 17:53:43 crc kubenswrapper[4808]: I1124 17:53:43.037935 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-wjrlj"] Nov 24 17:53:44 crc kubenswrapper[4808]: I1124 17:53:44.358132 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67ce2c95-e3aa-4d5f-8597-55c293b9e22b" path="/var/lib/kubelet/pods/67ce2c95-e3aa-4d5f-8597-55c293b9e22b/volumes" Nov 24 17:53:48 crc kubenswrapper[4808]: I1124 17:53:48.013051 4808 generic.go:334] "Generic (PLEG): container finished" podID="91323238-4f1e-4dd3-84f2-6fc8bf252642" containerID="92f2b8968dfc4e86050d852528d3c6677c69b6304cd5db0d86459b38da584526" exitCode=0 Nov 24 17:53:48 crc kubenswrapper[4808]: I1124 17:53:48.013209 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" event={"ID":"91323238-4f1e-4dd3-84f2-6fc8bf252642","Type":"ContainerDied","Data":"92f2b8968dfc4e86050d852528d3c6677c69b6304cd5db0d86459b38da584526"} Nov 24 17:53:48 crc kubenswrapper[4808]: I1124 17:53:48.358751 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:53:48 crc kubenswrapper[4808]: E1124 17:53:48.359561 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:53:49 crc kubenswrapper[4808]: I1124 17:53:49.394518 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" Nov 24 17:53:49 crc kubenswrapper[4808]: I1124 17:53:49.576871 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91323238-4f1e-4dd3-84f2-6fc8bf252642-ssh-key\") pod \"91323238-4f1e-4dd3-84f2-6fc8bf252642\" (UID: \"91323238-4f1e-4dd3-84f2-6fc8bf252642\") " Nov 24 17:53:49 crc kubenswrapper[4808]: I1124 17:53:49.577259 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91323238-4f1e-4dd3-84f2-6fc8bf252642-inventory\") pod \"91323238-4f1e-4dd3-84f2-6fc8bf252642\" (UID: \"91323238-4f1e-4dd3-84f2-6fc8bf252642\") " Nov 24 17:53:49 crc kubenswrapper[4808]: I1124 17:53:49.577377 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w62dt\" (UniqueName: \"kubernetes.io/projected/91323238-4f1e-4dd3-84f2-6fc8bf252642-kube-api-access-w62dt\") pod \"91323238-4f1e-4dd3-84f2-6fc8bf252642\" (UID: \"91323238-4f1e-4dd3-84f2-6fc8bf252642\") " Nov 24 17:53:49 crc kubenswrapper[4808]: I1124 17:53:49.582954 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91323238-4f1e-4dd3-84f2-6fc8bf252642-kube-api-access-w62dt" (OuterVolumeSpecName: "kube-api-access-w62dt") pod "91323238-4f1e-4dd3-84f2-6fc8bf252642" (UID: "91323238-4f1e-4dd3-84f2-6fc8bf252642"). InnerVolumeSpecName "kube-api-access-w62dt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:53:49 crc kubenswrapper[4808]: I1124 17:53:49.607251 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91323238-4f1e-4dd3-84f2-6fc8bf252642-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "91323238-4f1e-4dd3-84f2-6fc8bf252642" (UID: "91323238-4f1e-4dd3-84f2-6fc8bf252642"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:53:49 crc kubenswrapper[4808]: I1124 17:53:49.607254 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91323238-4f1e-4dd3-84f2-6fc8bf252642-inventory" (OuterVolumeSpecName: "inventory") pod "91323238-4f1e-4dd3-84f2-6fc8bf252642" (UID: "91323238-4f1e-4dd3-84f2-6fc8bf252642"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:53:49 crc kubenswrapper[4808]: I1124 17:53:49.679398 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w62dt\" (UniqueName: \"kubernetes.io/projected/91323238-4f1e-4dd3-84f2-6fc8bf252642-kube-api-access-w62dt\") on node \"crc\" DevicePath \"\"" Nov 24 17:53:49 crc kubenswrapper[4808]: I1124 17:53:49.679430 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91323238-4f1e-4dd3-84f2-6fc8bf252642-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:53:49 crc kubenswrapper[4808]: I1124 17:53:49.679440 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91323238-4f1e-4dd3-84f2-6fc8bf252642-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.026746 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-vnqqn"] Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.051905 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-vnqqn"] Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.055072 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" event={"ID":"91323238-4f1e-4dd3-84f2-6fc8bf252642","Type":"ContainerDied","Data":"a282b26fbd55afcdfc0b6336c684ec2ce64e0e334ac6c496d7728653b0a8b7c9"} Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.055114 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wglmk" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.055129 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a282b26fbd55afcdfc0b6336c684ec2ce64e0e334ac6c496d7728653b0a8b7c9" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.128547 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj"] Nov 24 17:53:50 crc kubenswrapper[4808]: E1124 17:53:50.128973 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91323238-4f1e-4dd3-84f2-6fc8bf252642" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.128992 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="91323238-4f1e-4dd3-84f2-6fc8bf252642" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.129220 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="91323238-4f1e-4dd3-84f2-6fc8bf252642" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.129871 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.134672 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.134729 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.134751 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.134766 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.142238 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj"] Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.289688 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzmqb\" (UniqueName: \"kubernetes.io/projected/579d95f1-e760-4476-bc98-6e33dabf6780-kube-api-access-tzmqb\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z92sj\" (UID: \"579d95f1-e760-4476-bc98-6e33dabf6780\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.290256 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/579d95f1-e760-4476-bc98-6e33dabf6780-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z92sj\" (UID: \"579d95f1-e760-4476-bc98-6e33dabf6780\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.290490 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/579d95f1-e760-4476-bc98-6e33dabf6780-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z92sj\" (UID: \"579d95f1-e760-4476-bc98-6e33dabf6780\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.372307 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71142c75-1dcb-4b06-92bd-f9882aa6a929" path="/var/lib/kubelet/pods/71142c75-1dcb-4b06-92bd-f9882aa6a929/volumes" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.393131 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzmqb\" (UniqueName: \"kubernetes.io/projected/579d95f1-e760-4476-bc98-6e33dabf6780-kube-api-access-tzmqb\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z92sj\" (UID: \"579d95f1-e760-4476-bc98-6e33dabf6780\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.393207 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/579d95f1-e760-4476-bc98-6e33dabf6780-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z92sj\" (UID: \"579d95f1-e760-4476-bc98-6e33dabf6780\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.393305 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/579d95f1-e760-4476-bc98-6e33dabf6780-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z92sj\" (UID: \"579d95f1-e760-4476-bc98-6e33dabf6780\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.399635 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/579d95f1-e760-4476-bc98-6e33dabf6780-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z92sj\" (UID: \"579d95f1-e760-4476-bc98-6e33dabf6780\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.400140 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/579d95f1-e760-4476-bc98-6e33dabf6780-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z92sj\" (UID: \"579d95f1-e760-4476-bc98-6e33dabf6780\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.415282 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzmqb\" (UniqueName: \"kubernetes.io/projected/579d95f1-e760-4476-bc98-6e33dabf6780-kube-api-access-tzmqb\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z92sj\" (UID: \"579d95f1-e760-4476-bc98-6e33dabf6780\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" Nov 24 17:53:50 crc kubenswrapper[4808]: I1124 17:53:50.456413 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" Nov 24 17:53:51 crc kubenswrapper[4808]: I1124 17:53:51.020888 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj"] Nov 24 17:53:51 crc kubenswrapper[4808]: W1124 17:53:51.023852 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod579d95f1_e760_4476_bc98_6e33dabf6780.slice/crio-a484139fc0bdf71486faf27250804c9aa58b8788f449793b9098573f5e3ea7b6 WatchSource:0}: Error finding container a484139fc0bdf71486faf27250804c9aa58b8788f449793b9098573f5e3ea7b6: Status 404 returned error can't find the container with id a484139fc0bdf71486faf27250804c9aa58b8788f449793b9098573f5e3ea7b6 Nov 24 17:53:51 crc kubenswrapper[4808]: I1124 17:53:51.064126 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" event={"ID":"579d95f1-e760-4476-bc98-6e33dabf6780","Type":"ContainerStarted","Data":"a484139fc0bdf71486faf27250804c9aa58b8788f449793b9098573f5e3ea7b6"} Nov 24 17:53:52 crc kubenswrapper[4808]: I1124 17:53:52.073611 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" event={"ID":"579d95f1-e760-4476-bc98-6e33dabf6780","Type":"ContainerStarted","Data":"f59a109046930e2e029f5633cd4c652e295dd921284195cb3625161d5b0b1419"} Nov 24 17:53:52 crc kubenswrapper[4808]: I1124 17:53:52.100967 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" podStartSLOduration=1.695366753 podStartE2EDuration="2.100935861s" podCreationTimestamp="2025-11-24 17:53:50 +0000 UTC" firstStartedPulling="2025-11-24 17:53:51.025694436 +0000 UTC m=+1623.623362238" lastFinishedPulling="2025-11-24 17:53:51.431263544 +0000 UTC m=+1624.028931346" observedRunningTime="2025-11-24 17:53:52.094425719 +0000 UTC m=+1624.692093521" watchObservedRunningTime="2025-11-24 17:53:52.100935861 +0000 UTC m=+1624.698603683" Nov 24 17:53:53 crc kubenswrapper[4808]: I1124 17:53:53.274095 4808 scope.go:117] "RemoveContainer" containerID="425eda6db936f103a37c96001c727487212eb86f3674f8faeb79054849f1e6ba" Nov 24 17:53:53 crc kubenswrapper[4808]: I1124 17:53:53.325476 4808 scope.go:117] "RemoveContainer" containerID="6187b82d0d321402af6fcc5c3a2a3b802f357dff62d1fc526ebb3c3fce434725" Nov 24 17:53:53 crc kubenswrapper[4808]: I1124 17:53:53.364640 4808 scope.go:117] "RemoveContainer" containerID="afa0a9324c77e5234ce460beb02ca8d9614289ad4644a4244b1a739e7c1072f4" Nov 24 17:53:53 crc kubenswrapper[4808]: I1124 17:53:53.402965 4808 scope.go:117] "RemoveContainer" containerID="0f300ffe4c22dc83f9d9f75dc05b4ddfb6d512f92c83fd1be7ea1ee284a8a765" Nov 24 17:53:53 crc kubenswrapper[4808]: I1124 17:53:53.454902 4808 scope.go:117] "RemoveContainer" containerID="b91696c1a899298ddd07d20470b57336c1acdd027253fcf40687a2cb526de70c" Nov 24 17:53:57 crc kubenswrapper[4808]: I1124 17:53:57.121125 4808 generic.go:334] "Generic (PLEG): container finished" podID="579d95f1-e760-4476-bc98-6e33dabf6780" containerID="f59a109046930e2e029f5633cd4c652e295dd921284195cb3625161d5b0b1419" exitCode=0 Nov 24 17:53:57 crc kubenswrapper[4808]: I1124 17:53:57.121175 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" event={"ID":"579d95f1-e760-4476-bc98-6e33dabf6780","Type":"ContainerDied","Data":"f59a109046930e2e029f5633cd4c652e295dd921284195cb3625161d5b0b1419"} Nov 24 17:53:58 crc kubenswrapper[4808]: I1124 17:53:58.040774 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-p77zx"] Nov 24 17:53:58 crc kubenswrapper[4808]: I1124 17:53:58.050737 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-p77zx"] Nov 24 17:53:58 crc kubenswrapper[4808]: I1124 17:53:58.358766 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b80d86dd-fbd8-444a-8159-d5c870eb01be" path="/var/lib/kubelet/pods/b80d86dd-fbd8-444a-8159-d5c870eb01be/volumes" Nov 24 17:53:58 crc kubenswrapper[4808]: I1124 17:53:58.465779 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" Nov 24 17:53:58 crc kubenswrapper[4808]: I1124 17:53:58.549533 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/579d95f1-e760-4476-bc98-6e33dabf6780-ssh-key\") pod \"579d95f1-e760-4476-bc98-6e33dabf6780\" (UID: \"579d95f1-e760-4476-bc98-6e33dabf6780\") " Nov 24 17:53:58 crc kubenswrapper[4808]: I1124 17:53:58.549574 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzmqb\" (UniqueName: \"kubernetes.io/projected/579d95f1-e760-4476-bc98-6e33dabf6780-kube-api-access-tzmqb\") pod \"579d95f1-e760-4476-bc98-6e33dabf6780\" (UID: \"579d95f1-e760-4476-bc98-6e33dabf6780\") " Nov 24 17:53:58 crc kubenswrapper[4808]: I1124 17:53:58.549689 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/579d95f1-e760-4476-bc98-6e33dabf6780-inventory\") pod \"579d95f1-e760-4476-bc98-6e33dabf6780\" (UID: \"579d95f1-e760-4476-bc98-6e33dabf6780\") " Nov 24 17:53:58 crc kubenswrapper[4808]: I1124 17:53:58.555085 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/579d95f1-e760-4476-bc98-6e33dabf6780-kube-api-access-tzmqb" (OuterVolumeSpecName: "kube-api-access-tzmqb") pod "579d95f1-e760-4476-bc98-6e33dabf6780" (UID: "579d95f1-e760-4476-bc98-6e33dabf6780"). InnerVolumeSpecName "kube-api-access-tzmqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:53:58 crc kubenswrapper[4808]: I1124 17:53:58.579490 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/579d95f1-e760-4476-bc98-6e33dabf6780-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "579d95f1-e760-4476-bc98-6e33dabf6780" (UID: "579d95f1-e760-4476-bc98-6e33dabf6780"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:53:58 crc kubenswrapper[4808]: I1124 17:53:58.580177 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/579d95f1-e760-4476-bc98-6e33dabf6780-inventory" (OuterVolumeSpecName: "inventory") pod "579d95f1-e760-4476-bc98-6e33dabf6780" (UID: "579d95f1-e760-4476-bc98-6e33dabf6780"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:53:58 crc kubenswrapper[4808]: I1124 17:53:58.652112 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/579d95f1-e760-4476-bc98-6e33dabf6780-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:53:58 crc kubenswrapper[4808]: I1124 17:53:58.652426 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzmqb\" (UniqueName: \"kubernetes.io/projected/579d95f1-e760-4476-bc98-6e33dabf6780-kube-api-access-tzmqb\") on node \"crc\" DevicePath \"\"" Nov 24 17:53:58 crc kubenswrapper[4808]: I1124 17:53:58.652493 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/579d95f1-e760-4476-bc98-6e33dabf6780-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.146446 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" event={"ID":"579d95f1-e760-4476-bc98-6e33dabf6780","Type":"ContainerDied","Data":"a484139fc0bdf71486faf27250804c9aa58b8788f449793b9098573f5e3ea7b6"} Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.146737 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a484139fc0bdf71486faf27250804c9aa58b8788f449793b9098573f5e3ea7b6" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.146620 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z92sj" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.208893 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9"] Nov 24 17:53:59 crc kubenswrapper[4808]: E1124 17:53:59.209653 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="579d95f1-e760-4476-bc98-6e33dabf6780" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.209673 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="579d95f1-e760-4476-bc98-6e33dabf6780" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.209913 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="579d95f1-e760-4476-bc98-6e33dabf6780" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.211130 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.213230 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.213354 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.213675 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.214345 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.227328 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9"] Nov 24 17:53:59 crc kubenswrapper[4808]: E1124 17:53:59.298239 4808 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod579d95f1_e760_4476_bc98_6e33dabf6780.slice\": RecentStats: unable to find data in memory cache]" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.372635 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37c9ab78-83a5-4432-8c14-b1e973891168-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-f8mg9\" (UID: \"37c9ab78-83a5-4432-8c14-b1e973891168\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.372785 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37c9ab78-83a5-4432-8c14-b1e973891168-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-f8mg9\" (UID: \"37c9ab78-83a5-4432-8c14-b1e973891168\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.372834 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w94j4\" (UniqueName: \"kubernetes.io/projected/37c9ab78-83a5-4432-8c14-b1e973891168-kube-api-access-w94j4\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-f8mg9\" (UID: \"37c9ab78-83a5-4432-8c14-b1e973891168\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.474656 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37c9ab78-83a5-4432-8c14-b1e973891168-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-f8mg9\" (UID: \"37c9ab78-83a5-4432-8c14-b1e973891168\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.474797 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37c9ab78-83a5-4432-8c14-b1e973891168-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-f8mg9\" (UID: \"37c9ab78-83a5-4432-8c14-b1e973891168\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.474834 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w94j4\" (UniqueName: \"kubernetes.io/projected/37c9ab78-83a5-4432-8c14-b1e973891168-kube-api-access-w94j4\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-f8mg9\" (UID: \"37c9ab78-83a5-4432-8c14-b1e973891168\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.479819 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37c9ab78-83a5-4432-8c14-b1e973891168-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-f8mg9\" (UID: \"37c9ab78-83a5-4432-8c14-b1e973891168\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.480637 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37c9ab78-83a5-4432-8c14-b1e973891168-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-f8mg9\" (UID: \"37c9ab78-83a5-4432-8c14-b1e973891168\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.497406 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w94j4\" (UniqueName: \"kubernetes.io/projected/37c9ab78-83a5-4432-8c14-b1e973891168-kube-api-access-w94j4\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-f8mg9\" (UID: \"37c9ab78-83a5-4432-8c14-b1e973891168\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" Nov 24 17:53:59 crc kubenswrapper[4808]: I1124 17:53:59.536995 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" Nov 24 17:54:00 crc kubenswrapper[4808]: I1124 17:54:00.039413 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9"] Nov 24 17:54:00 crc kubenswrapper[4808]: I1124 17:54:00.158892 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" event={"ID":"37c9ab78-83a5-4432-8c14-b1e973891168","Type":"ContainerStarted","Data":"2e51690ef8f70143345bcc8a69e9f3411af2b96d87d5a792efc640160e7a7fc2"} Nov 24 17:54:02 crc kubenswrapper[4808]: I1124 17:54:02.176632 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" event={"ID":"37c9ab78-83a5-4432-8c14-b1e973891168","Type":"ContainerStarted","Data":"7c60cfefbc294c3f40878702df74f8291fac7c68f14360fbbf0c7df8813a5b05"} Nov 24 17:54:02 crc kubenswrapper[4808]: I1124 17:54:02.197321 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" podStartSLOduration=2.28395557 podStartE2EDuration="3.197297601s" podCreationTimestamp="2025-11-24 17:53:59 +0000 UTC" firstStartedPulling="2025-11-24 17:54:00.039899921 +0000 UTC m=+1632.637567733" lastFinishedPulling="2025-11-24 17:54:00.953241962 +0000 UTC m=+1633.550909764" observedRunningTime="2025-11-24 17:54:02.193519109 +0000 UTC m=+1634.791186961" watchObservedRunningTime="2025-11-24 17:54:02.197297601 +0000 UTC m=+1634.794965403" Nov 24 17:54:02 crc kubenswrapper[4808]: I1124 17:54:02.351890 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:54:02 crc kubenswrapper[4808]: E1124 17:54:02.352426 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:54:17 crc kubenswrapper[4808]: I1124 17:54:17.347894 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:54:17 crc kubenswrapper[4808]: E1124 17:54:17.348675 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:54:30 crc kubenswrapper[4808]: I1124 17:54:30.347571 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:54:30 crc kubenswrapper[4808]: E1124 17:54:30.348594 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.035640 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-6nxpn"] Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.045042 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-cec5-account-create-update-vzfxt"] Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.074286 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-6nxpn"] Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.085405 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-cec5-account-create-update-vzfxt"] Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.588646 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dg2p9"] Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.590726 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.595052 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/778e1493-4bb7-4deb-8631-c19fbc08092c-utilities\") pod \"redhat-marketplace-dg2p9\" (UID: \"778e1493-4bb7-4deb-8631-c19fbc08092c\") " pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.595150 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n8sm\" (UniqueName: \"kubernetes.io/projected/778e1493-4bb7-4deb-8631-c19fbc08092c-kube-api-access-6n8sm\") pod \"redhat-marketplace-dg2p9\" (UID: \"778e1493-4bb7-4deb-8631-c19fbc08092c\") " pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.595173 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/778e1493-4bb7-4deb-8631-c19fbc08092c-catalog-content\") pod \"redhat-marketplace-dg2p9\" (UID: \"778e1493-4bb7-4deb-8631-c19fbc08092c\") " pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.595836 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dg2p9"] Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.696560 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n8sm\" (UniqueName: \"kubernetes.io/projected/778e1493-4bb7-4deb-8631-c19fbc08092c-kube-api-access-6n8sm\") pod \"redhat-marketplace-dg2p9\" (UID: \"778e1493-4bb7-4deb-8631-c19fbc08092c\") " pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.696628 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/778e1493-4bb7-4deb-8631-c19fbc08092c-catalog-content\") pod \"redhat-marketplace-dg2p9\" (UID: \"778e1493-4bb7-4deb-8631-c19fbc08092c\") " pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.696954 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/778e1493-4bb7-4deb-8631-c19fbc08092c-utilities\") pod \"redhat-marketplace-dg2p9\" (UID: \"778e1493-4bb7-4deb-8631-c19fbc08092c\") " pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.697123 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/778e1493-4bb7-4deb-8631-c19fbc08092c-catalog-content\") pod \"redhat-marketplace-dg2p9\" (UID: \"778e1493-4bb7-4deb-8631-c19fbc08092c\") " pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.697265 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/778e1493-4bb7-4deb-8631-c19fbc08092c-utilities\") pod \"redhat-marketplace-dg2p9\" (UID: \"778e1493-4bb7-4deb-8631-c19fbc08092c\") " pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.717675 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n8sm\" (UniqueName: \"kubernetes.io/projected/778e1493-4bb7-4deb-8631-c19fbc08092c-kube-api-access-6n8sm\") pod \"redhat-marketplace-dg2p9\" (UID: \"778e1493-4bb7-4deb-8631-c19fbc08092c\") " pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:35 crc kubenswrapper[4808]: I1124 17:54:35.912383 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.070088 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-4a1f-account-create-update-tp9xc"] Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.083053 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-dvnqp"] Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.093152 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-f761-account-create-update-dmmw8"] Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.104840 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-7czsx"] Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.111719 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-7czsx"] Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.133271 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-f761-account-create-update-dmmw8"] Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.146916 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-dvnqp"] Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.153821 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-4a1f-account-create-update-tp9xc"] Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.356968 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d" path="/var/lib/kubelet/pods/09ba5ae3-f9f7-4d05-a11b-0f7af4f89b3d/volumes" Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.357622 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a398e94-3f7a-48ec-815f-79c5a095c9c6" path="/var/lib/kubelet/pods/0a398e94-3f7a-48ec-815f-79c5a095c9c6/volumes" Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.358270 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b835d11-9481-4841-8a77-cc923fbb999b" path="/var/lib/kubelet/pods/5b835d11-9481-4841-8a77-cc923fbb999b/volumes" Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.358902 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="752ed614-a980-483f-a0ea-6fd150a6b694" path="/var/lib/kubelet/pods/752ed614-a980-483f-a0ea-6fd150a6b694/volumes" Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.360032 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c" path="/var/lib/kubelet/pods/76d13cdc-5ef8-4cd8-b4b9-d84c58b3965c/volumes" Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.360636 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87188566-f3cf-4cea-bdd1-28f3c48f5b37" path="/var/lib/kubelet/pods/87188566-f3cf-4cea-bdd1-28f3c48f5b37/volumes" Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.413049 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dg2p9"] Nov 24 17:54:36 crc kubenswrapper[4808]: I1124 17:54:36.486605 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg2p9" event={"ID":"778e1493-4bb7-4deb-8631-c19fbc08092c","Type":"ContainerStarted","Data":"aee65ae47406e28159b48136bb80fba1a2153d3c425991fe8e6b908cd4f39c41"} Nov 24 17:54:37 crc kubenswrapper[4808]: I1124 17:54:37.495839 4808 generic.go:334] "Generic (PLEG): container finished" podID="778e1493-4bb7-4deb-8631-c19fbc08092c" containerID="012dc5ecf04a51363ee849864103ef19824e7369a249a83967ed075940c3d132" exitCode=0 Nov 24 17:54:37 crc kubenswrapper[4808]: I1124 17:54:37.495887 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg2p9" event={"ID":"778e1493-4bb7-4deb-8631-c19fbc08092c","Type":"ContainerDied","Data":"012dc5ecf04a51363ee849864103ef19824e7369a249a83967ed075940c3d132"} Nov 24 17:54:38 crc kubenswrapper[4808]: I1124 17:54:38.506834 4808 generic.go:334] "Generic (PLEG): container finished" podID="778e1493-4bb7-4deb-8631-c19fbc08092c" containerID="5b39e543756a4e64484d71fc68a77fa675742ca7b35a8876c2208f73ef004418" exitCode=0 Nov 24 17:54:38 crc kubenswrapper[4808]: I1124 17:54:38.506993 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg2p9" event={"ID":"778e1493-4bb7-4deb-8631-c19fbc08092c","Type":"ContainerDied","Data":"5b39e543756a4e64484d71fc68a77fa675742ca7b35a8876c2208f73ef004418"} Nov 24 17:54:39 crc kubenswrapper[4808]: I1124 17:54:39.516369 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg2p9" event={"ID":"778e1493-4bb7-4deb-8631-c19fbc08092c","Type":"ContainerStarted","Data":"d6b7738e7d66d6543efe2d7988b7f06624e043e57e53ab282655d3a45916c725"} Nov 24 17:54:39 crc kubenswrapper[4808]: I1124 17:54:39.517719 4808 generic.go:334] "Generic (PLEG): container finished" podID="37c9ab78-83a5-4432-8c14-b1e973891168" containerID="7c60cfefbc294c3f40878702df74f8291fac7c68f14360fbbf0c7df8813a5b05" exitCode=0 Nov 24 17:54:39 crc kubenswrapper[4808]: I1124 17:54:39.517766 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" event={"ID":"37c9ab78-83a5-4432-8c14-b1e973891168","Type":"ContainerDied","Data":"7c60cfefbc294c3f40878702df74f8291fac7c68f14360fbbf0c7df8813a5b05"} Nov 24 17:54:39 crc kubenswrapper[4808]: I1124 17:54:39.536433 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dg2p9" podStartSLOduration=3.142163663 podStartE2EDuration="4.536412333s" podCreationTimestamp="2025-11-24 17:54:35 +0000 UTC" firstStartedPulling="2025-11-24 17:54:37.49749318 +0000 UTC m=+1670.095160982" lastFinishedPulling="2025-11-24 17:54:38.89174185 +0000 UTC m=+1671.489409652" observedRunningTime="2025-11-24 17:54:39.53168267 +0000 UTC m=+1672.129350502" watchObservedRunningTime="2025-11-24 17:54:39.536412333 +0000 UTC m=+1672.134080145" Nov 24 17:54:40 crc kubenswrapper[4808]: I1124 17:54:40.896510 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.007118 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37c9ab78-83a5-4432-8c14-b1e973891168-inventory\") pod \"37c9ab78-83a5-4432-8c14-b1e973891168\" (UID: \"37c9ab78-83a5-4432-8c14-b1e973891168\") " Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.007374 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w94j4\" (UniqueName: \"kubernetes.io/projected/37c9ab78-83a5-4432-8c14-b1e973891168-kube-api-access-w94j4\") pod \"37c9ab78-83a5-4432-8c14-b1e973891168\" (UID: \"37c9ab78-83a5-4432-8c14-b1e973891168\") " Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.007417 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37c9ab78-83a5-4432-8c14-b1e973891168-ssh-key\") pod \"37c9ab78-83a5-4432-8c14-b1e973891168\" (UID: \"37c9ab78-83a5-4432-8c14-b1e973891168\") " Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.013186 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37c9ab78-83a5-4432-8c14-b1e973891168-kube-api-access-w94j4" (OuterVolumeSpecName: "kube-api-access-w94j4") pod "37c9ab78-83a5-4432-8c14-b1e973891168" (UID: "37c9ab78-83a5-4432-8c14-b1e973891168"). InnerVolumeSpecName "kube-api-access-w94j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.042680 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37c9ab78-83a5-4432-8c14-b1e973891168-inventory" (OuterVolumeSpecName: "inventory") pod "37c9ab78-83a5-4432-8c14-b1e973891168" (UID: "37c9ab78-83a5-4432-8c14-b1e973891168"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.043167 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37c9ab78-83a5-4432-8c14-b1e973891168-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "37c9ab78-83a5-4432-8c14-b1e973891168" (UID: "37c9ab78-83a5-4432-8c14-b1e973891168"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.111428 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w94j4\" (UniqueName: \"kubernetes.io/projected/37c9ab78-83a5-4432-8c14-b1e973891168-kube-api-access-w94j4\") on node \"crc\" DevicePath \"\"" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.111471 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37c9ab78-83a5-4432-8c14-b1e973891168-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.111482 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37c9ab78-83a5-4432-8c14-b1e973891168-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.550226 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" event={"ID":"37c9ab78-83a5-4432-8c14-b1e973891168","Type":"ContainerDied","Data":"2e51690ef8f70143345bcc8a69e9f3411af2b96d87d5a792efc640160e7a7fc2"} Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.550271 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e51690ef8f70143345bcc8a69e9f3411af2b96d87d5a792efc640160e7a7fc2" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.550312 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-f8mg9" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.708575 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk"] Nov 24 17:54:41 crc kubenswrapper[4808]: E1124 17:54:41.708960 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37c9ab78-83a5-4432-8c14-b1e973891168" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.708979 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="37c9ab78-83a5-4432-8c14-b1e973891168" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.709175 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="37c9ab78-83a5-4432-8c14-b1e973891168" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.709729 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.714821 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.714894 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.714958 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.715160 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.731568 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk"] Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.825684 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faadd5b0-6e65-467e-b510-090ee77362b4-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk\" (UID: \"faadd5b0-6e65-467e-b510-090ee77362b4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.825945 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faadd5b0-6e65-467e-b510-090ee77362b4-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk\" (UID: \"faadd5b0-6e65-467e-b510-090ee77362b4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.826246 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4j2d\" (UniqueName: \"kubernetes.io/projected/faadd5b0-6e65-467e-b510-090ee77362b4-kube-api-access-k4j2d\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk\" (UID: \"faadd5b0-6e65-467e-b510-090ee77362b4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.928287 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faadd5b0-6e65-467e-b510-090ee77362b4-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk\" (UID: \"faadd5b0-6e65-467e-b510-090ee77362b4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.928333 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faadd5b0-6e65-467e-b510-090ee77362b4-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk\" (UID: \"faadd5b0-6e65-467e-b510-090ee77362b4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.928408 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4j2d\" (UniqueName: \"kubernetes.io/projected/faadd5b0-6e65-467e-b510-090ee77362b4-kube-api-access-k4j2d\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk\" (UID: \"faadd5b0-6e65-467e-b510-090ee77362b4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.933359 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faadd5b0-6e65-467e-b510-090ee77362b4-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk\" (UID: \"faadd5b0-6e65-467e-b510-090ee77362b4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.933770 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faadd5b0-6e65-467e-b510-090ee77362b4-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk\" (UID: \"faadd5b0-6e65-467e-b510-090ee77362b4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" Nov 24 17:54:41 crc kubenswrapper[4808]: I1124 17:54:41.946495 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4j2d\" (UniqueName: \"kubernetes.io/projected/faadd5b0-6e65-467e-b510-090ee77362b4-kube-api-access-k4j2d\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk\" (UID: \"faadd5b0-6e65-467e-b510-090ee77362b4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" Nov 24 17:54:42 crc kubenswrapper[4808]: I1124 17:54:42.030989 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" Nov 24 17:54:42 crc kubenswrapper[4808]: I1124 17:54:42.547102 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk"] Nov 24 17:54:42 crc kubenswrapper[4808]: W1124 17:54:42.551750 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfaadd5b0_6e65_467e_b510_090ee77362b4.slice/crio-7e6223ed99a1d5c50e4792a0ddd2618ad2a9a3350318a57013880e01318c6139 WatchSource:0}: Error finding container 7e6223ed99a1d5c50e4792a0ddd2618ad2a9a3350318a57013880e01318c6139: Status 404 returned error can't find the container with id 7e6223ed99a1d5c50e4792a0ddd2618ad2a9a3350318a57013880e01318c6139 Nov 24 17:54:43 crc kubenswrapper[4808]: I1124 17:54:43.569632 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" event={"ID":"faadd5b0-6e65-467e-b510-090ee77362b4","Type":"ContainerStarted","Data":"b058df246a51450c13002af11c046b63ba82633eeb090ca1576094e0bea1471f"} Nov 24 17:54:43 crc kubenswrapper[4808]: I1124 17:54:43.569876 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" event={"ID":"faadd5b0-6e65-467e-b510-090ee77362b4","Type":"ContainerStarted","Data":"7e6223ed99a1d5c50e4792a0ddd2618ad2a9a3350318a57013880e01318c6139"} Nov 24 17:54:43 crc kubenswrapper[4808]: I1124 17:54:43.591176 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" podStartSLOduration=2.041544021 podStartE2EDuration="2.591155114s" podCreationTimestamp="2025-11-24 17:54:41 +0000 UTC" firstStartedPulling="2025-11-24 17:54:42.554653672 +0000 UTC m=+1675.152321474" lastFinishedPulling="2025-11-24 17:54:43.104264765 +0000 UTC m=+1675.701932567" observedRunningTime="2025-11-24 17:54:43.586087953 +0000 UTC m=+1676.183755755" watchObservedRunningTime="2025-11-24 17:54:43.591155114 +0000 UTC m=+1676.188822926" Nov 24 17:54:44 crc kubenswrapper[4808]: I1124 17:54:44.347644 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:54:44 crc kubenswrapper[4808]: E1124 17:54:44.348150 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:54:45 crc kubenswrapper[4808]: I1124 17:54:45.912977 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:45 crc kubenswrapper[4808]: I1124 17:54:45.913092 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:45 crc kubenswrapper[4808]: I1124 17:54:45.959607 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:46 crc kubenswrapper[4808]: I1124 17:54:46.644179 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:46 crc kubenswrapper[4808]: I1124 17:54:46.703392 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dg2p9"] Nov 24 17:54:48 crc kubenswrapper[4808]: I1124 17:54:48.611288 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dg2p9" podUID="778e1493-4bb7-4deb-8631-c19fbc08092c" containerName="registry-server" containerID="cri-o://d6b7738e7d66d6543efe2d7988b7f06624e043e57e53ab282655d3a45916c725" gracePeriod=2 Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.560641 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.620080 4808 generic.go:334] "Generic (PLEG): container finished" podID="778e1493-4bb7-4deb-8631-c19fbc08092c" containerID="d6b7738e7d66d6543efe2d7988b7f06624e043e57e53ab282655d3a45916c725" exitCode=0 Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.620121 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dg2p9" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.620127 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg2p9" event={"ID":"778e1493-4bb7-4deb-8631-c19fbc08092c","Type":"ContainerDied","Data":"d6b7738e7d66d6543efe2d7988b7f06624e043e57e53ab282655d3a45916c725"} Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.620156 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg2p9" event={"ID":"778e1493-4bb7-4deb-8631-c19fbc08092c","Type":"ContainerDied","Data":"aee65ae47406e28159b48136bb80fba1a2153d3c425991fe8e6b908cd4f39c41"} Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.620177 4808 scope.go:117] "RemoveContainer" containerID="d6b7738e7d66d6543efe2d7988b7f06624e043e57e53ab282655d3a45916c725" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.641241 4808 scope.go:117] "RemoveContainer" containerID="5b39e543756a4e64484d71fc68a77fa675742ca7b35a8876c2208f73ef004418" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.664801 4808 scope.go:117] "RemoveContainer" containerID="012dc5ecf04a51363ee849864103ef19824e7369a249a83967ed075940c3d132" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.671269 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/778e1493-4bb7-4deb-8631-c19fbc08092c-catalog-content\") pod \"778e1493-4bb7-4deb-8631-c19fbc08092c\" (UID: \"778e1493-4bb7-4deb-8631-c19fbc08092c\") " Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.671409 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/778e1493-4bb7-4deb-8631-c19fbc08092c-utilities\") pod \"778e1493-4bb7-4deb-8631-c19fbc08092c\" (UID: \"778e1493-4bb7-4deb-8631-c19fbc08092c\") " Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.671513 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n8sm\" (UniqueName: \"kubernetes.io/projected/778e1493-4bb7-4deb-8631-c19fbc08092c-kube-api-access-6n8sm\") pod \"778e1493-4bb7-4deb-8631-c19fbc08092c\" (UID: \"778e1493-4bb7-4deb-8631-c19fbc08092c\") " Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.672574 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/778e1493-4bb7-4deb-8631-c19fbc08092c-utilities" (OuterVolumeSpecName: "utilities") pod "778e1493-4bb7-4deb-8631-c19fbc08092c" (UID: "778e1493-4bb7-4deb-8631-c19fbc08092c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.677641 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/778e1493-4bb7-4deb-8631-c19fbc08092c-kube-api-access-6n8sm" (OuterVolumeSpecName: "kube-api-access-6n8sm") pod "778e1493-4bb7-4deb-8631-c19fbc08092c" (UID: "778e1493-4bb7-4deb-8631-c19fbc08092c"). InnerVolumeSpecName "kube-api-access-6n8sm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.695766 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/778e1493-4bb7-4deb-8631-c19fbc08092c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "778e1493-4bb7-4deb-8631-c19fbc08092c" (UID: "778e1493-4bb7-4deb-8631-c19fbc08092c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.757707 4808 scope.go:117] "RemoveContainer" containerID="d6b7738e7d66d6543efe2d7988b7f06624e043e57e53ab282655d3a45916c725" Nov 24 17:54:49 crc kubenswrapper[4808]: E1124 17:54:49.758210 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6b7738e7d66d6543efe2d7988b7f06624e043e57e53ab282655d3a45916c725\": container with ID starting with d6b7738e7d66d6543efe2d7988b7f06624e043e57e53ab282655d3a45916c725 not found: ID does not exist" containerID="d6b7738e7d66d6543efe2d7988b7f06624e043e57e53ab282655d3a45916c725" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.758241 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6b7738e7d66d6543efe2d7988b7f06624e043e57e53ab282655d3a45916c725"} err="failed to get container status \"d6b7738e7d66d6543efe2d7988b7f06624e043e57e53ab282655d3a45916c725\": rpc error: code = NotFound desc = could not find container \"d6b7738e7d66d6543efe2d7988b7f06624e043e57e53ab282655d3a45916c725\": container with ID starting with d6b7738e7d66d6543efe2d7988b7f06624e043e57e53ab282655d3a45916c725 not found: ID does not exist" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.758267 4808 scope.go:117] "RemoveContainer" containerID="5b39e543756a4e64484d71fc68a77fa675742ca7b35a8876c2208f73ef004418" Nov 24 17:54:49 crc kubenswrapper[4808]: E1124 17:54:49.758556 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b39e543756a4e64484d71fc68a77fa675742ca7b35a8876c2208f73ef004418\": container with ID starting with 5b39e543756a4e64484d71fc68a77fa675742ca7b35a8876c2208f73ef004418 not found: ID does not exist" containerID="5b39e543756a4e64484d71fc68a77fa675742ca7b35a8876c2208f73ef004418" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.758594 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b39e543756a4e64484d71fc68a77fa675742ca7b35a8876c2208f73ef004418"} err="failed to get container status \"5b39e543756a4e64484d71fc68a77fa675742ca7b35a8876c2208f73ef004418\": rpc error: code = NotFound desc = could not find container \"5b39e543756a4e64484d71fc68a77fa675742ca7b35a8876c2208f73ef004418\": container with ID starting with 5b39e543756a4e64484d71fc68a77fa675742ca7b35a8876c2208f73ef004418 not found: ID does not exist" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.758615 4808 scope.go:117] "RemoveContainer" containerID="012dc5ecf04a51363ee849864103ef19824e7369a249a83967ed075940c3d132" Nov 24 17:54:49 crc kubenswrapper[4808]: E1124 17:54:49.759058 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"012dc5ecf04a51363ee849864103ef19824e7369a249a83967ed075940c3d132\": container with ID starting with 012dc5ecf04a51363ee849864103ef19824e7369a249a83967ed075940c3d132 not found: ID does not exist" containerID="012dc5ecf04a51363ee849864103ef19824e7369a249a83967ed075940c3d132" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.759094 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"012dc5ecf04a51363ee849864103ef19824e7369a249a83967ed075940c3d132"} err="failed to get container status \"012dc5ecf04a51363ee849864103ef19824e7369a249a83967ed075940c3d132\": rpc error: code = NotFound desc = could not find container \"012dc5ecf04a51363ee849864103ef19824e7369a249a83967ed075940c3d132\": container with ID starting with 012dc5ecf04a51363ee849864103ef19824e7369a249a83967ed075940c3d132 not found: ID does not exist" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.773482 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/778e1493-4bb7-4deb-8631-c19fbc08092c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.773530 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/778e1493-4bb7-4deb-8631-c19fbc08092c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.773544 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n8sm\" (UniqueName: \"kubernetes.io/projected/778e1493-4bb7-4deb-8631-c19fbc08092c-kube-api-access-6n8sm\") on node \"crc\" DevicePath \"\"" Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.960166 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dg2p9"] Nov 24 17:54:49 crc kubenswrapper[4808]: I1124 17:54:49.970324 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dg2p9"] Nov 24 17:54:50 crc kubenswrapper[4808]: I1124 17:54:50.359494 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="778e1493-4bb7-4deb-8631-c19fbc08092c" path="/var/lib/kubelet/pods/778e1493-4bb7-4deb-8631-c19fbc08092c/volumes" Nov 24 17:54:53 crc kubenswrapper[4808]: I1124 17:54:53.551295 4808 scope.go:117] "RemoveContainer" containerID="31bad8802492f6658a7fae1e789c7fbf7adf4204d366956ca2ee46c406b4b28c" Nov 24 17:54:53 crc kubenswrapper[4808]: I1124 17:54:53.574223 4808 scope.go:117] "RemoveContainer" containerID="7f46ba68d7c49118fb0b987a7944bc491a47727d684d6ddf806bbe1d7562b2e6" Nov 24 17:54:53 crc kubenswrapper[4808]: I1124 17:54:53.615219 4808 scope.go:117] "RemoveContainer" containerID="54379e86098921409fbbfccf84f368209dc1839fd7b8b1db807667415cb6cb70" Nov 24 17:54:53 crc kubenswrapper[4808]: I1124 17:54:53.667029 4808 scope.go:117] "RemoveContainer" containerID="ae616192db09d828ccb1875aedcd69bcdfa2d4db65ac1018f3364bf456a209f3" Nov 24 17:54:53 crc kubenswrapper[4808]: I1124 17:54:53.706247 4808 scope.go:117] "RemoveContainer" containerID="45c5433e0229e830366205749ed6a0fd4d2ceccb76c0056cba81f64404a6cd3a" Nov 24 17:54:53 crc kubenswrapper[4808]: I1124 17:54:53.761883 4808 scope.go:117] "RemoveContainer" containerID="8172fc826925b3aeec00348cad371bdd3ffb29488b839c7612d3b7c5c861b9ca" Nov 24 17:54:53 crc kubenswrapper[4808]: I1124 17:54:53.788188 4808 scope.go:117] "RemoveContainer" containerID="b6f3277bc56acbbf522afb17fc0bf4ce8d53aed1e4d1a2fa1a244aa5d87517dc" Nov 24 17:54:59 crc kubenswrapper[4808]: I1124 17:54:59.347290 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:54:59 crc kubenswrapper[4808]: E1124 17:54:59.347943 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:55:04 crc kubenswrapper[4808]: I1124 17:55:04.035138 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xz9xq"] Nov 24 17:55:04 crc kubenswrapper[4808]: I1124 17:55:04.044714 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xz9xq"] Nov 24 17:55:04 crc kubenswrapper[4808]: I1124 17:55:04.360804 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd6dfdb4-e6c9-4302-bdf5-cbb81348a510" path="/var/lib/kubelet/pods/bd6dfdb4-e6c9-4302-bdf5-cbb81348a510/volumes" Nov 24 17:55:12 crc kubenswrapper[4808]: I1124 17:55:12.348425 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:55:12 crc kubenswrapper[4808]: E1124 17:55:12.349774 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:55:27 crc kubenswrapper[4808]: I1124 17:55:27.041663 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-lfpxw"] Nov 24 17:55:27 crc kubenswrapper[4808]: I1124 17:55:27.049370 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-lfpxw"] Nov 24 17:55:27 crc kubenswrapper[4808]: I1124 17:55:27.347441 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:55:27 crc kubenswrapper[4808]: E1124 17:55:27.347721 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:55:28 crc kubenswrapper[4808]: I1124 17:55:28.358689 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ff6c89f-2920-4894-8036-bbf316ecd48b" path="/var/lib/kubelet/pods/1ff6c89f-2920-4894-8036-bbf316ecd48b/volumes" Nov 24 17:55:30 crc kubenswrapper[4808]: I1124 17:55:30.027613 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rfs5s"] Nov 24 17:55:30 crc kubenswrapper[4808]: I1124 17:55:30.036257 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rfs5s"] Nov 24 17:55:30 crc kubenswrapper[4808]: I1124 17:55:30.360880 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1373d3ee-b91b-413e-af94-40d6c55c1917" path="/var/lib/kubelet/pods/1373d3ee-b91b-413e-af94-40d6c55c1917/volumes" Nov 24 17:55:33 crc kubenswrapper[4808]: I1124 17:55:33.022131 4808 generic.go:334] "Generic (PLEG): container finished" podID="faadd5b0-6e65-467e-b510-090ee77362b4" containerID="b058df246a51450c13002af11c046b63ba82633eeb090ca1576094e0bea1471f" exitCode=0 Nov 24 17:55:33 crc kubenswrapper[4808]: I1124 17:55:33.022204 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" event={"ID":"faadd5b0-6e65-467e-b510-090ee77362b4","Type":"ContainerDied","Data":"b058df246a51450c13002af11c046b63ba82633eeb090ca1576094e0bea1471f"} Nov 24 17:55:34 crc kubenswrapper[4808]: I1124 17:55:34.454145 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" Nov 24 17:55:34 crc kubenswrapper[4808]: I1124 17:55:34.627777 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4j2d\" (UniqueName: \"kubernetes.io/projected/faadd5b0-6e65-467e-b510-090ee77362b4-kube-api-access-k4j2d\") pod \"faadd5b0-6e65-467e-b510-090ee77362b4\" (UID: \"faadd5b0-6e65-467e-b510-090ee77362b4\") " Nov 24 17:55:34 crc kubenswrapper[4808]: I1124 17:55:34.627933 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faadd5b0-6e65-467e-b510-090ee77362b4-ssh-key\") pod \"faadd5b0-6e65-467e-b510-090ee77362b4\" (UID: \"faadd5b0-6e65-467e-b510-090ee77362b4\") " Nov 24 17:55:34 crc kubenswrapper[4808]: I1124 17:55:34.628009 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faadd5b0-6e65-467e-b510-090ee77362b4-inventory\") pod \"faadd5b0-6e65-467e-b510-090ee77362b4\" (UID: \"faadd5b0-6e65-467e-b510-090ee77362b4\") " Nov 24 17:55:34 crc kubenswrapper[4808]: I1124 17:55:34.638353 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faadd5b0-6e65-467e-b510-090ee77362b4-kube-api-access-k4j2d" (OuterVolumeSpecName: "kube-api-access-k4j2d") pod "faadd5b0-6e65-467e-b510-090ee77362b4" (UID: "faadd5b0-6e65-467e-b510-090ee77362b4"). InnerVolumeSpecName "kube-api-access-k4j2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:55:34 crc kubenswrapper[4808]: I1124 17:55:34.656789 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faadd5b0-6e65-467e-b510-090ee77362b4-inventory" (OuterVolumeSpecName: "inventory") pod "faadd5b0-6e65-467e-b510-090ee77362b4" (UID: "faadd5b0-6e65-467e-b510-090ee77362b4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:55:34 crc kubenswrapper[4808]: I1124 17:55:34.673477 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faadd5b0-6e65-467e-b510-090ee77362b4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "faadd5b0-6e65-467e-b510-090ee77362b4" (UID: "faadd5b0-6e65-467e-b510-090ee77362b4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:55:34 crc kubenswrapper[4808]: I1124 17:55:34.730691 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faadd5b0-6e65-467e-b510-090ee77362b4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:34 crc kubenswrapper[4808]: I1124 17:55:34.730721 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faadd5b0-6e65-467e-b510-090ee77362b4-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:34 crc kubenswrapper[4808]: I1124 17:55:34.730732 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4j2d\" (UniqueName: \"kubernetes.io/projected/faadd5b0-6e65-467e-b510-090ee77362b4-kube-api-access-k4j2d\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.044371 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" event={"ID":"faadd5b0-6e65-467e-b510-090ee77362b4","Type":"ContainerDied","Data":"7e6223ed99a1d5c50e4792a0ddd2618ad2a9a3350318a57013880e01318c6139"} Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.044665 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e6223ed99a1d5c50e4792a0ddd2618ad2a9a3350318a57013880e01318c6139" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.044436 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.137148 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h2xqn"] Nov 24 17:55:35 crc kubenswrapper[4808]: E1124 17:55:35.137562 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778e1493-4bb7-4deb-8631-c19fbc08092c" containerName="registry-server" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.137587 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="778e1493-4bb7-4deb-8631-c19fbc08092c" containerName="registry-server" Nov 24 17:55:35 crc kubenswrapper[4808]: E1124 17:55:35.137614 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778e1493-4bb7-4deb-8631-c19fbc08092c" containerName="extract-utilities" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.137621 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="778e1493-4bb7-4deb-8631-c19fbc08092c" containerName="extract-utilities" Nov 24 17:55:35 crc kubenswrapper[4808]: E1124 17:55:35.137632 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faadd5b0-6e65-467e-b510-090ee77362b4" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.137640 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="faadd5b0-6e65-467e-b510-090ee77362b4" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:55:35 crc kubenswrapper[4808]: E1124 17:55:35.137655 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778e1493-4bb7-4deb-8631-c19fbc08092c" containerName="extract-content" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.137661 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="778e1493-4bb7-4deb-8631-c19fbc08092c" containerName="extract-content" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.137833 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="778e1493-4bb7-4deb-8631-c19fbc08092c" containerName="registry-server" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.137851 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="faadd5b0-6e65-467e-b510-090ee77362b4" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.138653 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.140958 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.141168 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.141431 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.142423 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.148987 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h2xqn"] Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.238149 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6bab5c2c-7363-40d3-855e-de98d478a482-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-h2xqn\" (UID: \"6bab5c2c-7363-40d3-855e-de98d478a482\") " pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.238226 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6bab5c2c-7363-40d3-855e-de98d478a482-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-h2xqn\" (UID: \"6bab5c2c-7363-40d3-855e-de98d478a482\") " pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.238399 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77nq2\" (UniqueName: \"kubernetes.io/projected/6bab5c2c-7363-40d3-855e-de98d478a482-kube-api-access-77nq2\") pod \"ssh-known-hosts-edpm-deployment-h2xqn\" (UID: \"6bab5c2c-7363-40d3-855e-de98d478a482\") " pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.340275 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6bab5c2c-7363-40d3-855e-de98d478a482-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-h2xqn\" (UID: \"6bab5c2c-7363-40d3-855e-de98d478a482\") " pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.340381 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6bab5c2c-7363-40d3-855e-de98d478a482-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-h2xqn\" (UID: \"6bab5c2c-7363-40d3-855e-de98d478a482\") " pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.340424 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77nq2\" (UniqueName: \"kubernetes.io/projected/6bab5c2c-7363-40d3-855e-de98d478a482-kube-api-access-77nq2\") pod \"ssh-known-hosts-edpm-deployment-h2xqn\" (UID: \"6bab5c2c-7363-40d3-855e-de98d478a482\") " pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.346721 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6bab5c2c-7363-40d3-855e-de98d478a482-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-h2xqn\" (UID: \"6bab5c2c-7363-40d3-855e-de98d478a482\") " pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.348350 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6bab5c2c-7363-40d3-855e-de98d478a482-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-h2xqn\" (UID: \"6bab5c2c-7363-40d3-855e-de98d478a482\") " pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.359784 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77nq2\" (UniqueName: \"kubernetes.io/projected/6bab5c2c-7363-40d3-855e-de98d478a482-kube-api-access-77nq2\") pod \"ssh-known-hosts-edpm-deployment-h2xqn\" (UID: \"6bab5c2c-7363-40d3-855e-de98d478a482\") " pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" Nov 24 17:55:35 crc kubenswrapper[4808]: I1124 17:55:35.462573 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" Nov 24 17:55:36 crc kubenswrapper[4808]: I1124 17:55:36.027917 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h2xqn"] Nov 24 17:55:36 crc kubenswrapper[4808]: W1124 17:55:36.037937 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6bab5c2c_7363_40d3_855e_de98d478a482.slice/crio-83a88a11d361e6e52aec8552fdc9cb3a6a42a617161facb31b80d1d6aaae070c WatchSource:0}: Error finding container 83a88a11d361e6e52aec8552fdc9cb3a6a42a617161facb31b80d1d6aaae070c: Status 404 returned error can't find the container with id 83a88a11d361e6e52aec8552fdc9cb3a6a42a617161facb31b80d1d6aaae070c Nov 24 17:55:36 crc kubenswrapper[4808]: I1124 17:55:36.052932 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" event={"ID":"6bab5c2c-7363-40d3-855e-de98d478a482","Type":"ContainerStarted","Data":"83a88a11d361e6e52aec8552fdc9cb3a6a42a617161facb31b80d1d6aaae070c"} Nov 24 17:55:37 crc kubenswrapper[4808]: I1124 17:55:37.062721 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" event={"ID":"6bab5c2c-7363-40d3-855e-de98d478a482","Type":"ContainerStarted","Data":"110c7cbe56fd1d4b1c51dcabd059c12edcb91f1040eadd20980d51884584d9bf"} Nov 24 17:55:37 crc kubenswrapper[4808]: I1124 17:55:37.091479 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" podStartSLOduration=1.667052567 podStartE2EDuration="2.091461396s" podCreationTimestamp="2025-11-24 17:55:35 +0000 UTC" firstStartedPulling="2025-11-24 17:55:36.040352256 +0000 UTC m=+1728.638020058" lastFinishedPulling="2025-11-24 17:55:36.464761065 +0000 UTC m=+1729.062428887" observedRunningTime="2025-11-24 17:55:37.079522997 +0000 UTC m=+1729.677190809" watchObservedRunningTime="2025-11-24 17:55:37.091461396 +0000 UTC m=+1729.689129198" Nov 24 17:55:42 crc kubenswrapper[4808]: I1124 17:55:42.348050 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:55:42 crc kubenswrapper[4808]: E1124 17:55:42.348963 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:55:44 crc kubenswrapper[4808]: I1124 17:55:44.122427 4808 generic.go:334] "Generic (PLEG): container finished" podID="6bab5c2c-7363-40d3-855e-de98d478a482" containerID="110c7cbe56fd1d4b1c51dcabd059c12edcb91f1040eadd20980d51884584d9bf" exitCode=0 Nov 24 17:55:44 crc kubenswrapper[4808]: I1124 17:55:44.122466 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" event={"ID":"6bab5c2c-7363-40d3-855e-de98d478a482","Type":"ContainerDied","Data":"110c7cbe56fd1d4b1c51dcabd059c12edcb91f1040eadd20980d51884584d9bf"} Nov 24 17:55:45 crc kubenswrapper[4808]: I1124 17:55:45.509056 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" Nov 24 17:55:45 crc kubenswrapper[4808]: I1124 17:55:45.626881 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6bab5c2c-7363-40d3-855e-de98d478a482-ssh-key-openstack-edpm-ipam\") pod \"6bab5c2c-7363-40d3-855e-de98d478a482\" (UID: \"6bab5c2c-7363-40d3-855e-de98d478a482\") " Nov 24 17:55:45 crc kubenswrapper[4808]: I1124 17:55:45.627043 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77nq2\" (UniqueName: \"kubernetes.io/projected/6bab5c2c-7363-40d3-855e-de98d478a482-kube-api-access-77nq2\") pod \"6bab5c2c-7363-40d3-855e-de98d478a482\" (UID: \"6bab5c2c-7363-40d3-855e-de98d478a482\") " Nov 24 17:55:45 crc kubenswrapper[4808]: I1124 17:55:45.627191 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6bab5c2c-7363-40d3-855e-de98d478a482-inventory-0\") pod \"6bab5c2c-7363-40d3-855e-de98d478a482\" (UID: \"6bab5c2c-7363-40d3-855e-de98d478a482\") " Nov 24 17:55:45 crc kubenswrapper[4808]: I1124 17:55:45.632643 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bab5c2c-7363-40d3-855e-de98d478a482-kube-api-access-77nq2" (OuterVolumeSpecName: "kube-api-access-77nq2") pod "6bab5c2c-7363-40d3-855e-de98d478a482" (UID: "6bab5c2c-7363-40d3-855e-de98d478a482"). InnerVolumeSpecName "kube-api-access-77nq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:55:45 crc kubenswrapper[4808]: I1124 17:55:45.654633 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bab5c2c-7363-40d3-855e-de98d478a482-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "6bab5c2c-7363-40d3-855e-de98d478a482" (UID: "6bab5c2c-7363-40d3-855e-de98d478a482"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:55:45 crc kubenswrapper[4808]: I1124 17:55:45.654687 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bab5c2c-7363-40d3-855e-de98d478a482-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6bab5c2c-7363-40d3-855e-de98d478a482" (UID: "6bab5c2c-7363-40d3-855e-de98d478a482"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:55:45 crc kubenswrapper[4808]: I1124 17:55:45.729943 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77nq2\" (UniqueName: \"kubernetes.io/projected/6bab5c2c-7363-40d3-855e-de98d478a482-kube-api-access-77nq2\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:45 crc kubenswrapper[4808]: I1124 17:55:45.730433 4808 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6bab5c2c-7363-40d3-855e-de98d478a482-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:45 crc kubenswrapper[4808]: I1124 17:55:45.730452 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6bab5c2c-7363-40d3-855e-de98d478a482-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.141765 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" event={"ID":"6bab5c2c-7363-40d3-855e-de98d478a482","Type":"ContainerDied","Data":"83a88a11d361e6e52aec8552fdc9cb3a6a42a617161facb31b80d1d6aaae070c"} Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.141815 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83a88a11d361e6e52aec8552fdc9cb3a6a42a617161facb31b80d1d6aaae070c" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.141820 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h2xqn" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.202840 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt"] Nov 24 17:55:46 crc kubenswrapper[4808]: E1124 17:55:46.203721 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bab5c2c-7363-40d3-855e-de98d478a482" containerName="ssh-known-hosts-edpm-deployment" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.203883 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bab5c2c-7363-40d3-855e-de98d478a482" containerName="ssh-known-hosts-edpm-deployment" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.204403 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bab5c2c-7363-40d3-855e-de98d478a482" containerName="ssh-known-hosts-edpm-deployment" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.205715 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.208556 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.208936 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.209170 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.209324 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.214744 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt"] Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.341483 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpb42\" (UniqueName: \"kubernetes.io/projected/92d01f67-8a79-4e09-afad-bf04575f554b-kube-api-access-bpb42\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qztqt\" (UID: \"92d01f67-8a79-4e09-afad-bf04575f554b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.341598 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92d01f67-8a79-4e09-afad-bf04575f554b-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qztqt\" (UID: \"92d01f67-8a79-4e09-afad-bf04575f554b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.341655 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92d01f67-8a79-4e09-afad-bf04575f554b-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qztqt\" (UID: \"92d01f67-8a79-4e09-afad-bf04575f554b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.443872 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92d01f67-8a79-4e09-afad-bf04575f554b-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qztqt\" (UID: \"92d01f67-8a79-4e09-afad-bf04575f554b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.443963 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92d01f67-8a79-4e09-afad-bf04575f554b-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qztqt\" (UID: \"92d01f67-8a79-4e09-afad-bf04575f554b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.444136 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpb42\" (UniqueName: \"kubernetes.io/projected/92d01f67-8a79-4e09-afad-bf04575f554b-kube-api-access-bpb42\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qztqt\" (UID: \"92d01f67-8a79-4e09-afad-bf04575f554b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.448902 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92d01f67-8a79-4e09-afad-bf04575f554b-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qztqt\" (UID: \"92d01f67-8a79-4e09-afad-bf04575f554b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.449296 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92d01f67-8a79-4e09-afad-bf04575f554b-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qztqt\" (UID: \"92d01f67-8a79-4e09-afad-bf04575f554b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.459663 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpb42\" (UniqueName: \"kubernetes.io/projected/92d01f67-8a79-4e09-afad-bf04575f554b-kube-api-access-bpb42\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qztqt\" (UID: \"92d01f67-8a79-4e09-afad-bf04575f554b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" Nov 24 17:55:46 crc kubenswrapper[4808]: I1124 17:55:46.529813 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" Nov 24 17:55:47 crc kubenswrapper[4808]: I1124 17:55:47.075287 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt"] Nov 24 17:55:47 crc kubenswrapper[4808]: I1124 17:55:47.152171 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" event={"ID":"92d01f67-8a79-4e09-afad-bf04575f554b","Type":"ContainerStarted","Data":"e938c2f08ffa59c87b4c20688a1012d045cca8fccc2e91e4e19cd9e434a98bc3"} Nov 24 17:55:48 crc kubenswrapper[4808]: I1124 17:55:48.160212 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" event={"ID":"92d01f67-8a79-4e09-afad-bf04575f554b","Type":"ContainerStarted","Data":"c6a825630bc9bb8be3480cafbc3b4d3e394f85dd18ef26c3826cb140addec101"} Nov 24 17:55:48 crc kubenswrapper[4808]: I1124 17:55:48.180355 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" podStartSLOduration=1.756411154 podStartE2EDuration="2.180331181s" podCreationTimestamp="2025-11-24 17:55:46 +0000 UTC" firstStartedPulling="2025-11-24 17:55:47.08091688 +0000 UTC m=+1739.678584682" lastFinishedPulling="2025-11-24 17:55:47.504836907 +0000 UTC m=+1740.102504709" observedRunningTime="2025-11-24 17:55:48.173054763 +0000 UTC m=+1740.770722575" watchObservedRunningTime="2025-11-24 17:55:48.180331181 +0000 UTC m=+1740.777999023" Nov 24 17:55:53 crc kubenswrapper[4808]: I1124 17:55:53.927358 4808 scope.go:117] "RemoveContainer" containerID="865fc6bbbfe63477f72e6511cdf09da548421a81f1f2d025c582821f266b8dda" Nov 24 17:55:53 crc kubenswrapper[4808]: I1124 17:55:53.995330 4808 scope.go:117] "RemoveContainer" containerID="522929826dc3bfc36bdc89b2ee5bc6c4d9d79ae11afa1c70662522d45d6b603d" Nov 24 17:55:54 crc kubenswrapper[4808]: I1124 17:55:54.048534 4808 scope.go:117] "RemoveContainer" containerID="82ba366ba8d51289ea732089d13e7cadb0dd0c2c292e82a6b1e301f97a7ae5ab" Nov 24 17:55:55 crc kubenswrapper[4808]: I1124 17:55:55.233744 4808 generic.go:334] "Generic (PLEG): container finished" podID="92d01f67-8a79-4e09-afad-bf04575f554b" containerID="c6a825630bc9bb8be3480cafbc3b4d3e394f85dd18ef26c3826cb140addec101" exitCode=0 Nov 24 17:55:55 crc kubenswrapper[4808]: I1124 17:55:55.234026 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" event={"ID":"92d01f67-8a79-4e09-afad-bf04575f554b","Type":"ContainerDied","Data":"c6a825630bc9bb8be3480cafbc3b4d3e394f85dd18ef26c3826cb140addec101"} Nov 24 17:55:56 crc kubenswrapper[4808]: I1124 17:55:56.347453 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:55:56 crc kubenswrapper[4808]: E1124 17:55:56.348124 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:55:56 crc kubenswrapper[4808]: I1124 17:55:56.744348 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" Nov 24 17:55:56 crc kubenswrapper[4808]: I1124 17:55:56.853001 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpb42\" (UniqueName: \"kubernetes.io/projected/92d01f67-8a79-4e09-afad-bf04575f554b-kube-api-access-bpb42\") pod \"92d01f67-8a79-4e09-afad-bf04575f554b\" (UID: \"92d01f67-8a79-4e09-afad-bf04575f554b\") " Nov 24 17:55:56 crc kubenswrapper[4808]: I1124 17:55:56.853147 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92d01f67-8a79-4e09-afad-bf04575f554b-ssh-key\") pod \"92d01f67-8a79-4e09-afad-bf04575f554b\" (UID: \"92d01f67-8a79-4e09-afad-bf04575f554b\") " Nov 24 17:55:56 crc kubenswrapper[4808]: I1124 17:55:56.853253 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92d01f67-8a79-4e09-afad-bf04575f554b-inventory\") pod \"92d01f67-8a79-4e09-afad-bf04575f554b\" (UID: \"92d01f67-8a79-4e09-afad-bf04575f554b\") " Nov 24 17:55:56 crc kubenswrapper[4808]: I1124 17:55:56.858386 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d01f67-8a79-4e09-afad-bf04575f554b-kube-api-access-bpb42" (OuterVolumeSpecName: "kube-api-access-bpb42") pod "92d01f67-8a79-4e09-afad-bf04575f554b" (UID: "92d01f67-8a79-4e09-afad-bf04575f554b"). InnerVolumeSpecName "kube-api-access-bpb42". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:55:56 crc kubenswrapper[4808]: I1124 17:55:56.883774 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d01f67-8a79-4e09-afad-bf04575f554b-inventory" (OuterVolumeSpecName: "inventory") pod "92d01f67-8a79-4e09-afad-bf04575f554b" (UID: "92d01f67-8a79-4e09-afad-bf04575f554b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:55:56 crc kubenswrapper[4808]: I1124 17:55:56.887135 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d01f67-8a79-4e09-afad-bf04575f554b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "92d01f67-8a79-4e09-afad-bf04575f554b" (UID: "92d01f67-8a79-4e09-afad-bf04575f554b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:55:56 crc kubenswrapper[4808]: I1124 17:55:56.954999 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpb42\" (UniqueName: \"kubernetes.io/projected/92d01f67-8a79-4e09-afad-bf04575f554b-kube-api-access-bpb42\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:56 crc kubenswrapper[4808]: I1124 17:55:56.955059 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92d01f67-8a79-4e09-afad-bf04575f554b-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:56 crc kubenswrapper[4808]: I1124 17:55:56.955076 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92d01f67-8a79-4e09-afad-bf04575f554b-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.252553 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" event={"ID":"92d01f67-8a79-4e09-afad-bf04575f554b","Type":"ContainerDied","Data":"e938c2f08ffa59c87b4c20688a1012d045cca8fccc2e91e4e19cd9e434a98bc3"} Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.252893 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e938c2f08ffa59c87b4c20688a1012d045cca8fccc2e91e4e19cd9e434a98bc3" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.252629 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qztqt" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.375181 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s"] Nov 24 17:55:57 crc kubenswrapper[4808]: E1124 17:55:57.375519 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d01f67-8a79-4e09-afad-bf04575f554b" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.375531 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d01f67-8a79-4e09-afad-bf04575f554b" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.375722 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d01f67-8a79-4e09-afad-bf04575f554b" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.376435 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.379723 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.379954 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.380304 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.382559 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.391242 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s"] Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.463282 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a9a918d-83b1-4e41-89a1-15e375a5f18e-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s\" (UID: \"2a9a918d-83b1-4e41-89a1-15e375a5f18e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.463441 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x74mk\" (UniqueName: \"kubernetes.io/projected/2a9a918d-83b1-4e41-89a1-15e375a5f18e-kube-api-access-x74mk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s\" (UID: \"2a9a918d-83b1-4e41-89a1-15e375a5f18e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.463496 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a9a918d-83b1-4e41-89a1-15e375a5f18e-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s\" (UID: \"2a9a918d-83b1-4e41-89a1-15e375a5f18e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.565222 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a9a918d-83b1-4e41-89a1-15e375a5f18e-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s\" (UID: \"2a9a918d-83b1-4e41-89a1-15e375a5f18e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.565307 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x74mk\" (UniqueName: \"kubernetes.io/projected/2a9a918d-83b1-4e41-89a1-15e375a5f18e-kube-api-access-x74mk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s\" (UID: \"2a9a918d-83b1-4e41-89a1-15e375a5f18e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.565350 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a9a918d-83b1-4e41-89a1-15e375a5f18e-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s\" (UID: \"2a9a918d-83b1-4e41-89a1-15e375a5f18e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.569443 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a9a918d-83b1-4e41-89a1-15e375a5f18e-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s\" (UID: \"2a9a918d-83b1-4e41-89a1-15e375a5f18e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.569709 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a9a918d-83b1-4e41-89a1-15e375a5f18e-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s\" (UID: \"2a9a918d-83b1-4e41-89a1-15e375a5f18e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.588832 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x74mk\" (UniqueName: \"kubernetes.io/projected/2a9a918d-83b1-4e41-89a1-15e375a5f18e-kube-api-access-x74mk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s\" (UID: \"2a9a918d-83b1-4e41-89a1-15e375a5f18e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" Nov 24 17:55:57 crc kubenswrapper[4808]: I1124 17:55:57.699539 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" Nov 24 17:55:58 crc kubenswrapper[4808]: I1124 17:55:58.195920 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s"] Nov 24 17:55:58 crc kubenswrapper[4808]: I1124 17:55:58.264948 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" event={"ID":"2a9a918d-83b1-4e41-89a1-15e375a5f18e","Type":"ContainerStarted","Data":"9ebe69b2a2fe1c0c4237dd6f11945c936d72067b5461be7cdf432c0ea45835ca"} Nov 24 17:55:59 crc kubenswrapper[4808]: I1124 17:55:59.274465 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" event={"ID":"2a9a918d-83b1-4e41-89a1-15e375a5f18e","Type":"ContainerStarted","Data":"d1a4cb816543ea1533b713198373759ec7df112233f192540a511b4056cecfbf"} Nov 24 17:55:59 crc kubenswrapper[4808]: I1124 17:55:59.303313 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" podStartSLOduration=1.7790604669999999 podStartE2EDuration="2.303292911s" podCreationTimestamp="2025-11-24 17:55:57 +0000 UTC" firstStartedPulling="2025-11-24 17:55:58.209912318 +0000 UTC m=+1750.807580120" lastFinishedPulling="2025-11-24 17:55:58.734144762 +0000 UTC m=+1751.331812564" observedRunningTime="2025-11-24 17:55:59.291651498 +0000 UTC m=+1751.889319310" watchObservedRunningTime="2025-11-24 17:55:59.303292911 +0000 UTC m=+1751.900960713" Nov 24 17:56:08 crc kubenswrapper[4808]: I1124 17:56:08.356369 4808 generic.go:334] "Generic (PLEG): container finished" podID="2a9a918d-83b1-4e41-89a1-15e375a5f18e" containerID="d1a4cb816543ea1533b713198373759ec7df112233f192540a511b4056cecfbf" exitCode=0 Nov 24 17:56:08 crc kubenswrapper[4808]: I1124 17:56:08.360491 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" event={"ID":"2a9a918d-83b1-4e41-89a1-15e375a5f18e","Type":"ContainerDied","Data":"d1a4cb816543ea1533b713198373759ec7df112233f192540a511b4056cecfbf"} Nov 24 17:56:09 crc kubenswrapper[4808]: I1124 17:56:09.761139 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" Nov 24 17:56:09 crc kubenswrapper[4808]: I1124 17:56:09.819245 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a9a918d-83b1-4e41-89a1-15e375a5f18e-inventory\") pod \"2a9a918d-83b1-4e41-89a1-15e375a5f18e\" (UID: \"2a9a918d-83b1-4e41-89a1-15e375a5f18e\") " Nov 24 17:56:09 crc kubenswrapper[4808]: I1124 17:56:09.819293 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x74mk\" (UniqueName: \"kubernetes.io/projected/2a9a918d-83b1-4e41-89a1-15e375a5f18e-kube-api-access-x74mk\") pod \"2a9a918d-83b1-4e41-89a1-15e375a5f18e\" (UID: \"2a9a918d-83b1-4e41-89a1-15e375a5f18e\") " Nov 24 17:56:09 crc kubenswrapper[4808]: I1124 17:56:09.819353 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a9a918d-83b1-4e41-89a1-15e375a5f18e-ssh-key\") pod \"2a9a918d-83b1-4e41-89a1-15e375a5f18e\" (UID: \"2a9a918d-83b1-4e41-89a1-15e375a5f18e\") " Nov 24 17:56:09 crc kubenswrapper[4808]: I1124 17:56:09.825440 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a9a918d-83b1-4e41-89a1-15e375a5f18e-kube-api-access-x74mk" (OuterVolumeSpecName: "kube-api-access-x74mk") pod "2a9a918d-83b1-4e41-89a1-15e375a5f18e" (UID: "2a9a918d-83b1-4e41-89a1-15e375a5f18e"). InnerVolumeSpecName "kube-api-access-x74mk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:56:09 crc kubenswrapper[4808]: I1124 17:56:09.848390 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a9a918d-83b1-4e41-89a1-15e375a5f18e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2a9a918d-83b1-4e41-89a1-15e375a5f18e" (UID: "2a9a918d-83b1-4e41-89a1-15e375a5f18e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:56:09 crc kubenswrapper[4808]: I1124 17:56:09.852540 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a9a918d-83b1-4e41-89a1-15e375a5f18e-inventory" (OuterVolumeSpecName: "inventory") pod "2a9a918d-83b1-4e41-89a1-15e375a5f18e" (UID: "2a9a918d-83b1-4e41-89a1-15e375a5f18e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:56:09 crc kubenswrapper[4808]: I1124 17:56:09.921927 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a9a918d-83b1-4e41-89a1-15e375a5f18e-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:09 crc kubenswrapper[4808]: I1124 17:56:09.921966 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x74mk\" (UniqueName: \"kubernetes.io/projected/2a9a918d-83b1-4e41-89a1-15e375a5f18e-kube-api-access-x74mk\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:09 crc kubenswrapper[4808]: I1124 17:56:09.921976 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a9a918d-83b1-4e41-89a1-15e375a5f18e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.377141 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" event={"ID":"2a9a918d-83b1-4e41-89a1-15e375a5f18e","Type":"ContainerDied","Data":"9ebe69b2a2fe1c0c4237dd6f11945c936d72067b5461be7cdf432c0ea45835ca"} Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.377193 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ebe69b2a2fe1c0c4237dd6f11945c936d72067b5461be7cdf432c0ea45835ca" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.377253 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.505236 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf"] Nov 24 17:56:10 crc kubenswrapper[4808]: E1124 17:56:10.505659 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a9a918d-83b1-4e41-89a1-15e375a5f18e" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.505681 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a9a918d-83b1-4e41-89a1-15e375a5f18e" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.505899 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a9a918d-83b1-4e41-89a1-15e375a5f18e" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.506688 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.510421 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.510610 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.510764 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.511044 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.511170 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.511590 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.512216 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.518670 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf"] Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.518737 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.636281 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.636381 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.636441 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.636478 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.636514 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.636558 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.636615 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.636665 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.636729 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.636769 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.636844 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.636880 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.636926 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84t24\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-kube-api-access-84t24\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.637045 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.739262 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.739695 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.739770 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.739848 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.739910 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.739987 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.740103 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.740199 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.740285 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.740375 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.740441 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.740541 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.740596 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.740676 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84t24\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-kube-api-access-84t24\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.746945 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.747011 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.749087 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.749902 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.750263 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.750748 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.750924 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.751056 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.751086 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.751310 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.752959 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.756483 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.757775 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.759939 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84t24\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-kube-api-access-84t24\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:10 crc kubenswrapper[4808]: I1124 17:56:10.867610 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:11 crc kubenswrapper[4808]: I1124 17:56:11.152094 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf"] Nov 24 17:56:11 crc kubenswrapper[4808]: I1124 17:56:11.347121 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:56:11 crc kubenswrapper[4808]: E1124 17:56:11.347450 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:56:11 crc kubenswrapper[4808]: I1124 17:56:11.385770 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" event={"ID":"91e767d6-f4dd-4bf8-b58d-37dc9073e90f","Type":"ContainerStarted","Data":"9591518fe677975ecf2846bb4e61b40caef4db6591f87cc3eb31d7dbf5a7eaeb"} Nov 24 17:56:12 crc kubenswrapper[4808]: I1124 17:56:12.078218 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-6r66h"] Nov 24 17:56:12 crc kubenswrapper[4808]: I1124 17:56:12.092396 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-6r66h"] Nov 24 17:56:12 crc kubenswrapper[4808]: I1124 17:56:12.359582 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dac3443-b917-41b6-90c9-b23c01524462" path="/var/lib/kubelet/pods/3dac3443-b917-41b6-90c9-b23c01524462/volumes" Nov 24 17:56:12 crc kubenswrapper[4808]: I1124 17:56:12.396511 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" event={"ID":"91e767d6-f4dd-4bf8-b58d-37dc9073e90f","Type":"ContainerStarted","Data":"3dc41be6c0f6bcadbeb88fe649dd341c10e9d83abc652839316e9a4b97165ec5"} Nov 24 17:56:12 crc kubenswrapper[4808]: I1124 17:56:12.432808 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" podStartSLOduration=1.982527282 podStartE2EDuration="2.432791241s" podCreationTimestamp="2025-11-24 17:56:10 +0000 UTC" firstStartedPulling="2025-11-24 17:56:11.156447624 +0000 UTC m=+1763.754115426" lastFinishedPulling="2025-11-24 17:56:11.606711583 +0000 UTC m=+1764.204379385" observedRunningTime="2025-11-24 17:56:12.427773367 +0000 UTC m=+1765.025441169" watchObservedRunningTime="2025-11-24 17:56:12.432791241 +0000 UTC m=+1765.030459043" Nov 24 17:56:26 crc kubenswrapper[4808]: I1124 17:56:26.348106 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:56:26 crc kubenswrapper[4808]: E1124 17:56:26.349286 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:56:39 crc kubenswrapper[4808]: I1124 17:56:39.347068 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:56:39 crc kubenswrapper[4808]: E1124 17:56:39.347871 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:56:48 crc kubenswrapper[4808]: I1124 17:56:48.711012 4808 generic.go:334] "Generic (PLEG): container finished" podID="91e767d6-f4dd-4bf8-b58d-37dc9073e90f" containerID="3dc41be6c0f6bcadbeb88fe649dd341c10e9d83abc652839316e9a4b97165ec5" exitCode=0 Nov 24 17:56:48 crc kubenswrapper[4808]: I1124 17:56:48.711062 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" event={"ID":"91e767d6-f4dd-4bf8-b58d-37dc9073e90f","Type":"ContainerDied","Data":"3dc41be6c0f6bcadbeb88fe649dd341c10e9d83abc652839316e9a4b97165ec5"} Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.105162 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.199844 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84t24\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-kube-api-access-84t24\") pod \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.199905 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-neutron-metadata-combined-ca-bundle\") pod \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.200007 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.200107 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.200149 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-telemetry-combined-ca-bundle\") pod \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.200198 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-ovn-combined-ca-bundle\") pod \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.200307 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.200344 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.200380 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-bootstrap-combined-ca-bundle\") pod \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.200475 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-repo-setup-combined-ca-bundle\") pod \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.200519 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-ssh-key\") pod \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.200566 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-nova-combined-ca-bundle\") pod \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.200606 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-libvirt-combined-ca-bundle\") pod \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.200720 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-inventory\") pod \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\" (UID: \"91e767d6-f4dd-4bf8-b58d-37dc9073e90f\") " Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.207201 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "91e767d6-f4dd-4bf8-b58d-37dc9073e90f" (UID: "91e767d6-f4dd-4bf8-b58d-37dc9073e90f"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.208048 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "91e767d6-f4dd-4bf8-b58d-37dc9073e90f" (UID: "91e767d6-f4dd-4bf8-b58d-37dc9073e90f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.208147 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "91e767d6-f4dd-4bf8-b58d-37dc9073e90f" (UID: "91e767d6-f4dd-4bf8-b58d-37dc9073e90f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.208691 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "91e767d6-f4dd-4bf8-b58d-37dc9073e90f" (UID: "91e767d6-f4dd-4bf8-b58d-37dc9073e90f"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.209649 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "91e767d6-f4dd-4bf8-b58d-37dc9073e90f" (UID: "91e767d6-f4dd-4bf8-b58d-37dc9073e90f"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.209668 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-kube-api-access-84t24" (OuterVolumeSpecName: "kube-api-access-84t24") pod "91e767d6-f4dd-4bf8-b58d-37dc9073e90f" (UID: "91e767d6-f4dd-4bf8-b58d-37dc9073e90f"). InnerVolumeSpecName "kube-api-access-84t24". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.209786 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "91e767d6-f4dd-4bf8-b58d-37dc9073e90f" (UID: "91e767d6-f4dd-4bf8-b58d-37dc9073e90f"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.210069 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "91e767d6-f4dd-4bf8-b58d-37dc9073e90f" (UID: "91e767d6-f4dd-4bf8-b58d-37dc9073e90f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.211017 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "91e767d6-f4dd-4bf8-b58d-37dc9073e90f" (UID: "91e767d6-f4dd-4bf8-b58d-37dc9073e90f"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.211978 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "91e767d6-f4dd-4bf8-b58d-37dc9073e90f" (UID: "91e767d6-f4dd-4bf8-b58d-37dc9073e90f"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.212625 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "91e767d6-f4dd-4bf8-b58d-37dc9073e90f" (UID: "91e767d6-f4dd-4bf8-b58d-37dc9073e90f"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.221553 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "91e767d6-f4dd-4bf8-b58d-37dc9073e90f" (UID: "91e767d6-f4dd-4bf8-b58d-37dc9073e90f"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.234333 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "91e767d6-f4dd-4bf8-b58d-37dc9073e90f" (UID: "91e767d6-f4dd-4bf8-b58d-37dc9073e90f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.239489 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-inventory" (OuterVolumeSpecName: "inventory") pod "91e767d6-f4dd-4bf8-b58d-37dc9073e90f" (UID: "91e767d6-f4dd-4bf8-b58d-37dc9073e90f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.302192 4808 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.302226 4808 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.302239 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.302251 4808 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.302265 4808 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.302277 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.302289 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84t24\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-kube-api-access-84t24\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.302301 4808 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.302313 4808 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.302325 4808 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.302337 4808 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.302349 4808 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.302360 4808 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.302374 4808 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/91e767d6-f4dd-4bf8-b58d-37dc9073e90f-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.348412 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:56:50 crc kubenswrapper[4808]: E1124 17:56:50.348721 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.735992 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" event={"ID":"91e767d6-f4dd-4bf8-b58d-37dc9073e90f","Type":"ContainerDied","Data":"9591518fe677975ecf2846bb4e61b40caef4db6591f87cc3eb31d7dbf5a7eaeb"} Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.736082 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9591518fe677975ecf2846bb4e61b40caef4db6591f87cc3eb31d7dbf5a7eaeb" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.736105 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.848532 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8"] Nov 24 17:56:50 crc kubenswrapper[4808]: E1124 17:56:50.848967 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e767d6-f4dd-4bf8-b58d-37dc9073e90f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.848995 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e767d6-f4dd-4bf8-b58d-37dc9073e90f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.849243 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="91e767d6-f4dd-4bf8-b58d-37dc9073e90f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.849869 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.852879 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.852887 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.853686 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.854040 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.861401 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.864556 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8"] Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.913410 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5zl8\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.913496 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb6qv\" (UniqueName: \"kubernetes.io/projected/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-kube-api-access-cb6qv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5zl8\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.913554 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5zl8\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.913628 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5zl8\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:50 crc kubenswrapper[4808]: I1124 17:56:50.913680 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5zl8\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:51 crc kubenswrapper[4808]: I1124 17:56:51.015196 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb6qv\" (UniqueName: \"kubernetes.io/projected/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-kube-api-access-cb6qv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5zl8\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:51 crc kubenswrapper[4808]: I1124 17:56:51.015269 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5zl8\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:51 crc kubenswrapper[4808]: I1124 17:56:51.015313 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5zl8\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:51 crc kubenswrapper[4808]: I1124 17:56:51.015379 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5zl8\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:51 crc kubenswrapper[4808]: I1124 17:56:51.015438 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5zl8\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:51 crc kubenswrapper[4808]: I1124 17:56:51.016349 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5zl8\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:51 crc kubenswrapper[4808]: I1124 17:56:51.019522 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5zl8\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:51 crc kubenswrapper[4808]: I1124 17:56:51.019613 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5zl8\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:51 crc kubenswrapper[4808]: I1124 17:56:51.020552 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5zl8\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:51 crc kubenswrapper[4808]: I1124 17:56:51.032515 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb6qv\" (UniqueName: \"kubernetes.io/projected/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-kube-api-access-cb6qv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5zl8\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:51 crc kubenswrapper[4808]: I1124 17:56:51.178620 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:56:51 crc kubenswrapper[4808]: W1124 17:56:51.711562 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod119c4dcd_dc15_44aa_bfde_1a2931d8b83d.slice/crio-655144b12585d8e2778bfc28fb289a0953a45525a3d237fe8398d9af3482bbe7 WatchSource:0}: Error finding container 655144b12585d8e2778bfc28fb289a0953a45525a3d237fe8398d9af3482bbe7: Status 404 returned error can't find the container with id 655144b12585d8e2778bfc28fb289a0953a45525a3d237fe8398d9af3482bbe7 Nov 24 17:56:51 crc kubenswrapper[4808]: I1124 17:56:51.712791 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8"] Nov 24 17:56:51 crc kubenswrapper[4808]: I1124 17:56:51.748314 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" event={"ID":"119c4dcd-dc15-44aa-bfde-1a2931d8b83d","Type":"ContainerStarted","Data":"655144b12585d8e2778bfc28fb289a0953a45525a3d237fe8398d9af3482bbe7"} Nov 24 17:56:52 crc kubenswrapper[4808]: I1124 17:56:52.760508 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" event={"ID":"119c4dcd-dc15-44aa-bfde-1a2931d8b83d","Type":"ContainerStarted","Data":"0308eec27d09c8008e168410e7f8f63c80eda1c2624e129b6d8d8ba66df27b51"} Nov 24 17:56:52 crc kubenswrapper[4808]: I1124 17:56:52.789624 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" podStartSLOduration=2.215310151 podStartE2EDuration="2.789603347s" podCreationTimestamp="2025-11-24 17:56:50 +0000 UTC" firstStartedPulling="2025-11-24 17:56:51.714039463 +0000 UTC m=+1804.311707265" lastFinishedPulling="2025-11-24 17:56:52.288332639 +0000 UTC m=+1804.886000461" observedRunningTime="2025-11-24 17:56:52.783818901 +0000 UTC m=+1805.381486733" watchObservedRunningTime="2025-11-24 17:56:52.789603347 +0000 UTC m=+1805.387271159" Nov 24 17:56:54 crc kubenswrapper[4808]: I1124 17:56:54.225463 4808 scope.go:117] "RemoveContainer" containerID="eabf1eccc07ae41bc0daa8a260a65f4b49b7fe8db19eda8e0b5a97ad51312b1d" Nov 24 17:57:02 crc kubenswrapper[4808]: I1124 17:57:02.347419 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:57:02 crc kubenswrapper[4808]: E1124 17:57:02.348150 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:57:15 crc kubenswrapper[4808]: I1124 17:57:15.348175 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:57:15 crc kubenswrapper[4808]: E1124 17:57:15.349037 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:57:27 crc kubenswrapper[4808]: I1124 17:57:27.346867 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:57:27 crc kubenswrapper[4808]: E1124 17:57:27.347582 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:57:42 crc kubenswrapper[4808]: I1124 17:57:42.347653 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:57:42 crc kubenswrapper[4808]: E1124 17:57:42.348370 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:57:54 crc kubenswrapper[4808]: I1124 17:57:54.326415 4808 generic.go:334] "Generic (PLEG): container finished" podID="119c4dcd-dc15-44aa-bfde-1a2931d8b83d" containerID="0308eec27d09c8008e168410e7f8f63c80eda1c2624e129b6d8d8ba66df27b51" exitCode=0 Nov 24 17:57:54 crc kubenswrapper[4808]: I1124 17:57:54.326496 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" event={"ID":"119c4dcd-dc15-44aa-bfde-1a2931d8b83d","Type":"ContainerDied","Data":"0308eec27d09c8008e168410e7f8f63c80eda1c2624e129b6d8d8ba66df27b51"} Nov 24 17:57:55 crc kubenswrapper[4808]: I1124 17:57:55.347720 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:57:55 crc kubenswrapper[4808]: E1124 17:57:55.348358 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:57:55 crc kubenswrapper[4808]: I1124 17:57:55.775087 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:57:55 crc kubenswrapper[4808]: I1124 17:57:55.933106 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ovn-combined-ca-bundle\") pod \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " Nov 24 17:57:55 crc kubenswrapper[4808]: I1124 17:57:55.933190 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-inventory\") pod \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " Nov 24 17:57:55 crc kubenswrapper[4808]: I1124 17:57:55.933232 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ssh-key\") pod \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " Nov 24 17:57:55 crc kubenswrapper[4808]: I1124 17:57:55.933252 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb6qv\" (UniqueName: \"kubernetes.io/projected/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-kube-api-access-cb6qv\") pod \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " Nov 24 17:57:55 crc kubenswrapper[4808]: I1124 17:57:55.933414 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ovncontroller-config-0\") pod \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\" (UID: \"119c4dcd-dc15-44aa-bfde-1a2931d8b83d\") " Nov 24 17:57:55 crc kubenswrapper[4808]: I1124 17:57:55.938824 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "119c4dcd-dc15-44aa-bfde-1a2931d8b83d" (UID: "119c4dcd-dc15-44aa-bfde-1a2931d8b83d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:57:55 crc kubenswrapper[4808]: I1124 17:57:55.938837 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-kube-api-access-cb6qv" (OuterVolumeSpecName: "kube-api-access-cb6qv") pod "119c4dcd-dc15-44aa-bfde-1a2931d8b83d" (UID: "119c4dcd-dc15-44aa-bfde-1a2931d8b83d"). InnerVolumeSpecName "kube-api-access-cb6qv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:57:55 crc kubenswrapper[4808]: I1124 17:57:55.960152 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "119c4dcd-dc15-44aa-bfde-1a2931d8b83d" (UID: "119c4dcd-dc15-44aa-bfde-1a2931d8b83d"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:57:55 crc kubenswrapper[4808]: I1124 17:57:55.963224 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-inventory" (OuterVolumeSpecName: "inventory") pod "119c4dcd-dc15-44aa-bfde-1a2931d8b83d" (UID: "119c4dcd-dc15-44aa-bfde-1a2931d8b83d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:57:55 crc kubenswrapper[4808]: I1124 17:57:55.963263 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "119c4dcd-dc15-44aa-bfde-1a2931d8b83d" (UID: "119c4dcd-dc15-44aa-bfde-1a2931d8b83d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.036228 4808 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.036276 4808 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.036296 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.036309 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.036325 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb6qv\" (UniqueName: \"kubernetes.io/projected/119c4dcd-dc15-44aa-bfde-1a2931d8b83d-kube-api-access-cb6qv\") on node \"crc\" DevicePath \"\"" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.354914 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.375751 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5zl8" event={"ID":"119c4dcd-dc15-44aa-bfde-1a2931d8b83d","Type":"ContainerDied","Data":"655144b12585d8e2778bfc28fb289a0953a45525a3d237fe8398d9af3482bbe7"} Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.375806 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="655144b12585d8e2778bfc28fb289a0953a45525a3d237fe8398d9af3482bbe7" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.484616 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh"] Nov 24 17:57:56 crc kubenswrapper[4808]: E1124 17:57:56.485047 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="119c4dcd-dc15-44aa-bfde-1a2931d8b83d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.485067 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="119c4dcd-dc15-44aa-bfde-1a2931d8b83d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.485289 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="119c4dcd-dc15-44aa-bfde-1a2931d8b83d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.486171 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.489157 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.489280 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.489850 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.490495 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.490774 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.492050 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.497266 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh"] Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.647485 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrp9n\" (UniqueName: \"kubernetes.io/projected/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-kube-api-access-rrp9n\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.648138 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.648198 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.648307 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.648364 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.648389 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.751404 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.751586 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.751652 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.751811 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrp9n\" (UniqueName: \"kubernetes.io/projected/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-kube-api-access-rrp9n\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.751900 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.751959 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.757053 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.757264 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.757500 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.758255 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.759709 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.781834 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrp9n\" (UniqueName: \"kubernetes.io/projected/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-kube-api-access-rrp9n\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:56 crc kubenswrapper[4808]: I1124 17:57:56.808678 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:57:57 crc kubenswrapper[4808]: I1124 17:57:57.195980 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh"] Nov 24 17:57:57 crc kubenswrapper[4808]: I1124 17:57:57.200802 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:57:57 crc kubenswrapper[4808]: I1124 17:57:57.364205 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" event={"ID":"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f","Type":"ContainerStarted","Data":"249ef1c99a0a2168e81644e3ac24fdede3d0378649b275d60a91dcc8ca882b37"} Nov 24 17:57:58 crc kubenswrapper[4808]: I1124 17:57:58.374467 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" event={"ID":"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f","Type":"ContainerStarted","Data":"90509da2c01cbf7abf045e928be8b65a4e0a761bc42c26fba6ebc11fe4c9c47c"} Nov 24 17:57:58 crc kubenswrapper[4808]: I1124 17:57:58.403383 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" podStartSLOduration=1.937961269 podStartE2EDuration="2.40336151s" podCreationTimestamp="2025-11-24 17:57:56 +0000 UTC" firstStartedPulling="2025-11-24 17:57:57.20060749 +0000 UTC m=+1869.798275292" lastFinishedPulling="2025-11-24 17:57:57.666007721 +0000 UTC m=+1870.263675533" observedRunningTime="2025-11-24 17:57:58.398884753 +0000 UTC m=+1870.996552565" watchObservedRunningTime="2025-11-24 17:57:58.40336151 +0000 UTC m=+1871.001029322" Nov 24 17:58:07 crc kubenswrapper[4808]: I1124 17:58:07.347488 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:58:07 crc kubenswrapper[4808]: E1124 17:58:07.348259 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:58:22 crc kubenswrapper[4808]: I1124 17:58:22.348167 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:58:22 crc kubenswrapper[4808]: E1124 17:58:22.349456 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:58:33 crc kubenswrapper[4808]: I1124 17:58:33.347397 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:58:33 crc kubenswrapper[4808]: E1124 17:58:33.348281 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 17:58:45 crc kubenswrapper[4808]: I1124 17:58:45.784528 4808 generic.go:334] "Generic (PLEG): container finished" podID="40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f" containerID="90509da2c01cbf7abf045e928be8b65a4e0a761bc42c26fba6ebc11fe4c9c47c" exitCode=0 Nov 24 17:58:45 crc kubenswrapper[4808]: I1124 17:58:45.784633 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" event={"ID":"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f","Type":"ContainerDied","Data":"90509da2c01cbf7abf045e928be8b65a4e0a761bc42c26fba6ebc11fe4c9c47c"} Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.222377 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.342150 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrp9n\" (UniqueName: \"kubernetes.io/projected/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-kube-api-access-rrp9n\") pod \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.342237 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-neutron-ovn-metadata-agent-neutron-config-0\") pod \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.342267 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-nova-metadata-neutron-config-0\") pod \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.342329 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-neutron-metadata-combined-ca-bundle\") pod \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.342387 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-ssh-key\") pod \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.342484 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-inventory\") pod \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\" (UID: \"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f\") " Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.347593 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-kube-api-access-rrp9n" (OuterVolumeSpecName: "kube-api-access-rrp9n") pod "40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f" (UID: "40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f"). InnerVolumeSpecName "kube-api-access-rrp9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.347797 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.348302 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f" (UID: "40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.385491 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f" (UID: "40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.389649 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f" (UID: "40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.397774 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f" (UID: "40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.399119 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-inventory" (OuterVolumeSpecName: "inventory") pod "40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f" (UID: "40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.449739 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrp9n\" (UniqueName: \"kubernetes.io/projected/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-kube-api-access-rrp9n\") on node \"crc\" DevicePath \"\"" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.449837 4808 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.449898 4808 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.449924 4808 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.449944 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.449996 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.805119 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" event={"ID":"40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f","Type":"ContainerDied","Data":"249ef1c99a0a2168e81644e3ac24fdede3d0378649b275d60a91dcc8ca882b37"} Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.805542 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="249ef1c99a0a2168e81644e3ac24fdede3d0378649b275d60a91dcc8ca882b37" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.805605 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.809385 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"a01acbe1d0e5b11ae5eb995ee52c951450e7a150c9c344c5eacfa7fe2bda2db2"} Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.948093 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2"] Nov 24 17:58:47 crc kubenswrapper[4808]: E1124 17:58:47.948869 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.948952 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.949248 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.950042 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.956765 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.956940 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.957106 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.957223 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.957223 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:58:47 crc kubenswrapper[4808]: I1124 17:58:47.962202 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2"] Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.060242 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.060366 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.060418 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sws4p\" (UniqueName: \"kubernetes.io/projected/3a26fcdc-f080-4677-a308-b08584153734-kube-api-access-sws4p\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.060446 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.060538 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.162913 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.163244 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.163478 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.163655 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sws4p\" (UniqueName: \"kubernetes.io/projected/3a26fcdc-f080-4677-a308-b08584153734-kube-api-access-sws4p\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.163795 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.170551 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.171367 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.174537 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.180676 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.182775 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sws4p\" (UniqueName: \"kubernetes.io/projected/3a26fcdc-f080-4677-a308-b08584153734-kube-api-access-sws4p\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.309825 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.318675 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 17:58:48 crc kubenswrapper[4808]: I1124 17:58:48.885928 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2"] Nov 24 17:58:48 crc kubenswrapper[4808]: W1124 17:58:48.890639 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a26fcdc_f080_4677_a308_b08584153734.slice/crio-55776be9c796f5071c8cc1bbc6e856dd7b96590762dac3d637952e25e8b4e3bf WatchSource:0}: Error finding container 55776be9c796f5071c8cc1bbc6e856dd7b96590762dac3d637952e25e8b4e3bf: Status 404 returned error can't find the container with id 55776be9c796f5071c8cc1bbc6e856dd7b96590762dac3d637952e25e8b4e3bf Nov 24 17:58:49 crc kubenswrapper[4808]: I1124 17:58:49.345971 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:58:49 crc kubenswrapper[4808]: I1124 17:58:49.829212 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" event={"ID":"3a26fcdc-f080-4677-a308-b08584153734","Type":"ContainerStarted","Data":"cad27741ed70c1921571017076dc2e7ee1b42883f999a6896d5ac43fa374246e"} Nov 24 17:58:49 crc kubenswrapper[4808]: I1124 17:58:49.829704 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" event={"ID":"3a26fcdc-f080-4677-a308-b08584153734","Type":"ContainerStarted","Data":"55776be9c796f5071c8cc1bbc6e856dd7b96590762dac3d637952e25e8b4e3bf"} Nov 24 17:58:49 crc kubenswrapper[4808]: I1124 17:58:49.846200 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" podStartSLOduration=2.40034011 podStartE2EDuration="2.84617808s" podCreationTimestamp="2025-11-24 17:58:47 +0000 UTC" firstStartedPulling="2025-11-24 17:58:48.896716917 +0000 UTC m=+1921.494384719" lastFinishedPulling="2025-11-24 17:58:49.342554887 +0000 UTC m=+1921.940222689" observedRunningTime="2025-11-24 17:58:49.841887151 +0000 UTC m=+1922.439554953" watchObservedRunningTime="2025-11-24 17:58:49.84617808 +0000 UTC m=+1922.443845882" Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.168558 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f"] Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.172078 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.174581 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.178904 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.182779 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f"] Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.306991 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/74e18530-914f-401e-a5ea-992db5d8785a-secret-volume\") pod \"collect-profiles-29400120-4dp4f\" (UID: \"74e18530-914f-401e-a5ea-992db5d8785a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.307099 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4jz2\" (UniqueName: \"kubernetes.io/projected/74e18530-914f-401e-a5ea-992db5d8785a-kube-api-access-z4jz2\") pod \"collect-profiles-29400120-4dp4f\" (UID: \"74e18530-914f-401e-a5ea-992db5d8785a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.307154 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/74e18530-914f-401e-a5ea-992db5d8785a-config-volume\") pod \"collect-profiles-29400120-4dp4f\" (UID: \"74e18530-914f-401e-a5ea-992db5d8785a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.409477 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/74e18530-914f-401e-a5ea-992db5d8785a-config-volume\") pod \"collect-profiles-29400120-4dp4f\" (UID: \"74e18530-914f-401e-a5ea-992db5d8785a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.409733 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/74e18530-914f-401e-a5ea-992db5d8785a-secret-volume\") pod \"collect-profiles-29400120-4dp4f\" (UID: \"74e18530-914f-401e-a5ea-992db5d8785a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.409861 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4jz2\" (UniqueName: \"kubernetes.io/projected/74e18530-914f-401e-a5ea-992db5d8785a-kube-api-access-z4jz2\") pod \"collect-profiles-29400120-4dp4f\" (UID: \"74e18530-914f-401e-a5ea-992db5d8785a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.410444 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/74e18530-914f-401e-a5ea-992db5d8785a-config-volume\") pod \"collect-profiles-29400120-4dp4f\" (UID: \"74e18530-914f-401e-a5ea-992db5d8785a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.429646 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/74e18530-914f-401e-a5ea-992db5d8785a-secret-volume\") pod \"collect-profiles-29400120-4dp4f\" (UID: \"74e18530-914f-401e-a5ea-992db5d8785a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.429645 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4jz2\" (UniqueName: \"kubernetes.io/projected/74e18530-914f-401e-a5ea-992db5d8785a-kube-api-access-z4jz2\") pod \"collect-profiles-29400120-4dp4f\" (UID: \"74e18530-914f-401e-a5ea-992db5d8785a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.492677 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" Nov 24 18:00:00 crc kubenswrapper[4808]: I1124 18:00:00.936990 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f"] Nov 24 18:00:01 crc kubenswrapper[4808]: I1124 18:00:01.465942 4808 generic.go:334] "Generic (PLEG): container finished" podID="74e18530-914f-401e-a5ea-992db5d8785a" containerID="62f38ac43fbb3d899ad2ef968c78ce3fcab171fd1356022bf71caf16e1fe93e8" exitCode=0 Nov 24 18:00:01 crc kubenswrapper[4808]: I1124 18:00:01.466004 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" event={"ID":"74e18530-914f-401e-a5ea-992db5d8785a","Type":"ContainerDied","Data":"62f38ac43fbb3d899ad2ef968c78ce3fcab171fd1356022bf71caf16e1fe93e8"} Nov 24 18:00:01 crc kubenswrapper[4808]: I1124 18:00:01.466276 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" event={"ID":"74e18530-914f-401e-a5ea-992db5d8785a","Type":"ContainerStarted","Data":"a730e7461b8dd7e97709d4131075606d8e0cd7748723961c207fde817f2e74d8"} Nov 24 18:00:02 crc kubenswrapper[4808]: I1124 18:00:02.826495 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" Nov 24 18:00:02 crc kubenswrapper[4808]: I1124 18:00:02.961888 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/74e18530-914f-401e-a5ea-992db5d8785a-secret-volume\") pod \"74e18530-914f-401e-a5ea-992db5d8785a\" (UID: \"74e18530-914f-401e-a5ea-992db5d8785a\") " Nov 24 18:00:02 crc kubenswrapper[4808]: I1124 18:00:02.961976 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4jz2\" (UniqueName: \"kubernetes.io/projected/74e18530-914f-401e-a5ea-992db5d8785a-kube-api-access-z4jz2\") pod \"74e18530-914f-401e-a5ea-992db5d8785a\" (UID: \"74e18530-914f-401e-a5ea-992db5d8785a\") " Nov 24 18:00:02 crc kubenswrapper[4808]: I1124 18:00:02.962056 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/74e18530-914f-401e-a5ea-992db5d8785a-config-volume\") pod \"74e18530-914f-401e-a5ea-992db5d8785a\" (UID: \"74e18530-914f-401e-a5ea-992db5d8785a\") " Nov 24 18:00:02 crc kubenswrapper[4808]: I1124 18:00:02.962961 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74e18530-914f-401e-a5ea-992db5d8785a-config-volume" (OuterVolumeSpecName: "config-volume") pod "74e18530-914f-401e-a5ea-992db5d8785a" (UID: "74e18530-914f-401e-a5ea-992db5d8785a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 18:00:02 crc kubenswrapper[4808]: I1124 18:00:02.967419 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74e18530-914f-401e-a5ea-992db5d8785a-kube-api-access-z4jz2" (OuterVolumeSpecName: "kube-api-access-z4jz2") pod "74e18530-914f-401e-a5ea-992db5d8785a" (UID: "74e18530-914f-401e-a5ea-992db5d8785a"). InnerVolumeSpecName "kube-api-access-z4jz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:00:02 crc kubenswrapper[4808]: I1124 18:00:02.967862 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74e18530-914f-401e-a5ea-992db5d8785a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "74e18530-914f-401e-a5ea-992db5d8785a" (UID: "74e18530-914f-401e-a5ea-992db5d8785a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:00:03 crc kubenswrapper[4808]: I1124 18:00:03.065003 4808 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/74e18530-914f-401e-a5ea-992db5d8785a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 18:00:03 crc kubenswrapper[4808]: I1124 18:00:03.065281 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4jz2\" (UniqueName: \"kubernetes.io/projected/74e18530-914f-401e-a5ea-992db5d8785a-kube-api-access-z4jz2\") on node \"crc\" DevicePath \"\"" Nov 24 18:00:03 crc kubenswrapper[4808]: I1124 18:00:03.065292 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/74e18530-914f-401e-a5ea-992db5d8785a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 18:00:03 crc kubenswrapper[4808]: I1124 18:00:03.486305 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" event={"ID":"74e18530-914f-401e-a5ea-992db5d8785a","Type":"ContainerDied","Data":"a730e7461b8dd7e97709d4131075606d8e0cd7748723961c207fde817f2e74d8"} Nov 24 18:00:03 crc kubenswrapper[4808]: I1124 18:00:03.486342 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a730e7461b8dd7e97709d4131075606d8e0cd7748723961c207fde817f2e74d8" Nov 24 18:00:03 crc kubenswrapper[4808]: I1124 18:00:03.486777 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4dp4f" Nov 24 18:00:03 crc kubenswrapper[4808]: I1124 18:00:03.905484 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn"] Nov 24 18:00:03 crc kubenswrapper[4808]: I1124 18:00:03.912072 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400075-dhldn"] Nov 24 18:00:04 crc kubenswrapper[4808]: I1124 18:00:04.358908 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0d60e28-fafc-4490-802e-ee578094c9a3" path="/var/lib/kubelet/pods/d0d60e28-fafc-4490-802e-ee578094c9a3/volumes" Nov 24 18:00:52 crc kubenswrapper[4808]: I1124 18:00:52.837470 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n7vbp"] Nov 24 18:00:52 crc kubenswrapper[4808]: E1124 18:00:52.838533 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74e18530-914f-401e-a5ea-992db5d8785a" containerName="collect-profiles" Nov 24 18:00:52 crc kubenswrapper[4808]: I1124 18:00:52.838552 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="74e18530-914f-401e-a5ea-992db5d8785a" containerName="collect-profiles" Nov 24 18:00:52 crc kubenswrapper[4808]: I1124 18:00:52.838782 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="74e18530-914f-401e-a5ea-992db5d8785a" containerName="collect-profiles" Nov 24 18:00:52 crc kubenswrapper[4808]: I1124 18:00:52.852795 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n7vbp"] Nov 24 18:00:52 crc kubenswrapper[4808]: I1124 18:00:52.852924 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:00:53 crc kubenswrapper[4808]: I1124 18:00:53.000190 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs84r\" (UniqueName: \"kubernetes.io/projected/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-kube-api-access-vs84r\") pod \"redhat-operators-n7vbp\" (UID: \"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5\") " pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:00:53 crc kubenswrapper[4808]: I1124 18:00:53.000300 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-catalog-content\") pod \"redhat-operators-n7vbp\" (UID: \"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5\") " pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:00:53 crc kubenswrapper[4808]: I1124 18:00:53.000338 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-utilities\") pod \"redhat-operators-n7vbp\" (UID: \"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5\") " pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:00:53 crc kubenswrapper[4808]: I1124 18:00:53.102353 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs84r\" (UniqueName: \"kubernetes.io/projected/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-kube-api-access-vs84r\") pod \"redhat-operators-n7vbp\" (UID: \"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5\") " pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:00:53 crc kubenswrapper[4808]: I1124 18:00:53.102436 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-catalog-content\") pod \"redhat-operators-n7vbp\" (UID: \"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5\") " pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:00:53 crc kubenswrapper[4808]: I1124 18:00:53.102469 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-utilities\") pod \"redhat-operators-n7vbp\" (UID: \"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5\") " pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:00:53 crc kubenswrapper[4808]: I1124 18:00:53.102984 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-utilities\") pod \"redhat-operators-n7vbp\" (UID: \"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5\") " pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:00:53 crc kubenswrapper[4808]: I1124 18:00:53.103163 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-catalog-content\") pod \"redhat-operators-n7vbp\" (UID: \"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5\") " pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:00:53 crc kubenswrapper[4808]: I1124 18:00:53.124813 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs84r\" (UniqueName: \"kubernetes.io/projected/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-kube-api-access-vs84r\") pod \"redhat-operators-n7vbp\" (UID: \"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5\") " pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:00:53 crc kubenswrapper[4808]: I1124 18:00:53.201871 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:00:53 crc kubenswrapper[4808]: I1124 18:00:53.689519 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n7vbp"] Nov 24 18:00:53 crc kubenswrapper[4808]: I1124 18:00:53.944185 4808 generic.go:334] "Generic (PLEG): container finished" podID="47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" containerID="5e6e66cad2aa309536c44c69b95e67bab23173965451f11d09b574d046457fc5" exitCode=0 Nov 24 18:00:53 crc kubenswrapper[4808]: I1124 18:00:53.944282 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7vbp" event={"ID":"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5","Type":"ContainerDied","Data":"5e6e66cad2aa309536c44c69b95e67bab23173965451f11d09b574d046457fc5"} Nov 24 18:00:53 crc kubenswrapper[4808]: I1124 18:00:53.944487 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7vbp" event={"ID":"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5","Type":"ContainerStarted","Data":"d40a4d838348c7ef643881da1ba5eddd57d99f931bba474833c1352f901d1923"} Nov 24 18:00:54 crc kubenswrapper[4808]: I1124 18:00:54.356556 4808 scope.go:117] "RemoveContainer" containerID="19e5c8ff96219b9fb2f9b99395f0296e89569dc4a1d45586980af9ef69cdc61a" Nov 24 18:00:55 crc kubenswrapper[4808]: I1124 18:00:55.967406 4808 generic.go:334] "Generic (PLEG): container finished" podID="47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" containerID="0284776c8a628f219687045d70ec892dcea3ef840d71b8d59931376d99674306" exitCode=0 Nov 24 18:00:55 crc kubenswrapper[4808]: I1124 18:00:55.967554 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7vbp" event={"ID":"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5","Type":"ContainerDied","Data":"0284776c8a628f219687045d70ec892dcea3ef840d71b8d59931376d99674306"} Nov 24 18:00:56 crc kubenswrapper[4808]: I1124 18:00:56.982140 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7vbp" event={"ID":"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5","Type":"ContainerStarted","Data":"1257e3df9f7dc2f362757f75ac9e7c7b38ba6dfd70510b308bfa4cca23493d32"} Nov 24 18:00:56 crc kubenswrapper[4808]: I1124 18:00:56.998218 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n7vbp" podStartSLOduration=2.50664305 podStartE2EDuration="4.998200019s" podCreationTimestamp="2025-11-24 18:00:52 +0000 UTC" firstStartedPulling="2025-11-24 18:00:53.945922534 +0000 UTC m=+2046.543590336" lastFinishedPulling="2025-11-24 18:00:56.437479493 +0000 UTC m=+2049.035147305" observedRunningTime="2025-11-24 18:00:56.997875521 +0000 UTC m=+2049.595543323" watchObservedRunningTime="2025-11-24 18:00:56.998200019 +0000 UTC m=+2049.595867821" Nov 24 18:00:59 crc kubenswrapper[4808]: I1124 18:00:59.612833 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lm2cj"] Nov 24 18:00:59 crc kubenswrapper[4808]: I1124 18:00:59.616348 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:00:59 crc kubenswrapper[4808]: I1124 18:00:59.621312 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lm2cj"] Nov 24 18:00:59 crc kubenswrapper[4808]: I1124 18:00:59.656980 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc2adcf3-c591-4319-80b4-737e5aa6d117-utilities\") pod \"community-operators-lm2cj\" (UID: \"cc2adcf3-c591-4319-80b4-737e5aa6d117\") " pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:00:59 crc kubenswrapper[4808]: I1124 18:00:59.657266 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc2adcf3-c591-4319-80b4-737e5aa6d117-catalog-content\") pod \"community-operators-lm2cj\" (UID: \"cc2adcf3-c591-4319-80b4-737e5aa6d117\") " pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:00:59 crc kubenswrapper[4808]: I1124 18:00:59.657324 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8hlt\" (UniqueName: \"kubernetes.io/projected/cc2adcf3-c591-4319-80b4-737e5aa6d117-kube-api-access-p8hlt\") pod \"community-operators-lm2cj\" (UID: \"cc2adcf3-c591-4319-80b4-737e5aa6d117\") " pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:00:59 crc kubenswrapper[4808]: I1124 18:00:59.758696 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc2adcf3-c591-4319-80b4-737e5aa6d117-utilities\") pod \"community-operators-lm2cj\" (UID: \"cc2adcf3-c591-4319-80b4-737e5aa6d117\") " pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:00:59 crc kubenswrapper[4808]: I1124 18:00:59.758878 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc2adcf3-c591-4319-80b4-737e5aa6d117-catalog-content\") pod \"community-operators-lm2cj\" (UID: \"cc2adcf3-c591-4319-80b4-737e5aa6d117\") " pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:00:59 crc kubenswrapper[4808]: I1124 18:00:59.758918 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8hlt\" (UniqueName: \"kubernetes.io/projected/cc2adcf3-c591-4319-80b4-737e5aa6d117-kube-api-access-p8hlt\") pod \"community-operators-lm2cj\" (UID: \"cc2adcf3-c591-4319-80b4-737e5aa6d117\") " pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:00:59 crc kubenswrapper[4808]: I1124 18:00:59.759248 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc2adcf3-c591-4319-80b4-737e5aa6d117-utilities\") pod \"community-operators-lm2cj\" (UID: \"cc2adcf3-c591-4319-80b4-737e5aa6d117\") " pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:00:59 crc kubenswrapper[4808]: I1124 18:00:59.759578 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc2adcf3-c591-4319-80b4-737e5aa6d117-catalog-content\") pod \"community-operators-lm2cj\" (UID: \"cc2adcf3-c591-4319-80b4-737e5aa6d117\") " pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:00:59 crc kubenswrapper[4808]: I1124 18:00:59.779129 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8hlt\" (UniqueName: \"kubernetes.io/projected/cc2adcf3-c591-4319-80b4-737e5aa6d117-kube-api-access-p8hlt\") pod \"community-operators-lm2cj\" (UID: \"cc2adcf3-c591-4319-80b4-737e5aa6d117\") " pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:00:59 crc kubenswrapper[4808]: I1124 18:00:59.974513 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.190138 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29400121-9z8sf"] Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.191620 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.217065 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29400121-9z8sf"] Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.379262 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-combined-ca-bundle\") pod \"keystone-cron-29400121-9z8sf\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.379350 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-fernet-keys\") pod \"keystone-cron-29400121-9z8sf\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.379377 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjh5g\" (UniqueName: \"kubernetes.io/projected/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-kube-api-access-mjh5g\") pod \"keystone-cron-29400121-9z8sf\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.379633 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-config-data\") pod \"keystone-cron-29400121-9z8sf\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.481637 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-config-data\") pod \"keystone-cron-29400121-9z8sf\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.481748 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-combined-ca-bundle\") pod \"keystone-cron-29400121-9z8sf\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.481787 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-fernet-keys\") pod \"keystone-cron-29400121-9z8sf\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.481802 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjh5g\" (UniqueName: \"kubernetes.io/projected/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-kube-api-access-mjh5g\") pod \"keystone-cron-29400121-9z8sf\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.488111 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-config-data\") pod \"keystone-cron-29400121-9z8sf\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.488128 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-fernet-keys\") pod \"keystone-cron-29400121-9z8sf\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.488409 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-combined-ca-bundle\") pod \"keystone-cron-29400121-9z8sf\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.504053 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjh5g\" (UniqueName: \"kubernetes.io/projected/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-kube-api-access-mjh5g\") pod \"keystone-cron-29400121-9z8sf\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.521428 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:00 crc kubenswrapper[4808]: I1124 18:01:00.539042 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lm2cj"] Nov 24 18:01:01 crc kubenswrapper[4808]: W1124 18:01:01.009096 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7d529f4_bce4_4d3a_83b2_38c65e0f4e7e.slice/crio-af4400ffa0bf4591d6ffbba3d634eb56c688ff6296e5276ffb00b68e1e494f88 WatchSource:0}: Error finding container af4400ffa0bf4591d6ffbba3d634eb56c688ff6296e5276ffb00b68e1e494f88: Status 404 returned error can't find the container with id af4400ffa0bf4591d6ffbba3d634eb56c688ff6296e5276ffb00b68e1e494f88 Nov 24 18:01:01 crc kubenswrapper[4808]: I1124 18:01:01.016508 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29400121-9z8sf"] Nov 24 18:01:01 crc kubenswrapper[4808]: I1124 18:01:01.030407 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400121-9z8sf" event={"ID":"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e","Type":"ContainerStarted","Data":"af4400ffa0bf4591d6ffbba3d634eb56c688ff6296e5276ffb00b68e1e494f88"} Nov 24 18:01:01 crc kubenswrapper[4808]: I1124 18:01:01.034135 4808 generic.go:334] "Generic (PLEG): container finished" podID="cc2adcf3-c591-4319-80b4-737e5aa6d117" containerID="f455ba8c1cc08576628feec2c770f6e2736f851cc891c7183a758d63e4ce32d1" exitCode=0 Nov 24 18:01:01 crc kubenswrapper[4808]: I1124 18:01:01.034198 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lm2cj" event={"ID":"cc2adcf3-c591-4319-80b4-737e5aa6d117","Type":"ContainerDied","Data":"f455ba8c1cc08576628feec2c770f6e2736f851cc891c7183a758d63e4ce32d1"} Nov 24 18:01:01 crc kubenswrapper[4808]: I1124 18:01:01.034261 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lm2cj" event={"ID":"cc2adcf3-c591-4319-80b4-737e5aa6d117","Type":"ContainerStarted","Data":"9acdacbf43bca14fead7907019a0180ac3061fac52f457410196835d2f70b897"} Nov 24 18:01:02 crc kubenswrapper[4808]: I1124 18:01:02.048463 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lm2cj" event={"ID":"cc2adcf3-c591-4319-80b4-737e5aa6d117","Type":"ContainerStarted","Data":"43e875974de5b9356acd3c791a09c7ea9e40d4356903ed15b98d894163684036"} Nov 24 18:01:02 crc kubenswrapper[4808]: I1124 18:01:02.051690 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400121-9z8sf" event={"ID":"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e","Type":"ContainerStarted","Data":"3d4f141ef589b731e5cb42b570b8bee6520343882b3d89a7221f8b54779f2079"} Nov 24 18:01:02 crc kubenswrapper[4808]: I1124 18:01:02.098909 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29400121-9z8sf" podStartSLOduration=2.098887053 podStartE2EDuration="2.098887053s" podCreationTimestamp="2025-11-24 18:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 18:01:02.092252201 +0000 UTC m=+2054.689920033" watchObservedRunningTime="2025-11-24 18:01:02.098887053 +0000 UTC m=+2054.696554875" Nov 24 18:01:03 crc kubenswrapper[4808]: I1124 18:01:03.064828 4808 generic.go:334] "Generic (PLEG): container finished" podID="cc2adcf3-c591-4319-80b4-737e5aa6d117" containerID="43e875974de5b9356acd3c791a09c7ea9e40d4356903ed15b98d894163684036" exitCode=0 Nov 24 18:01:03 crc kubenswrapper[4808]: I1124 18:01:03.067661 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lm2cj" event={"ID":"cc2adcf3-c591-4319-80b4-737e5aa6d117","Type":"ContainerDied","Data":"43e875974de5b9356acd3c791a09c7ea9e40d4356903ed15b98d894163684036"} Nov 24 18:01:03 crc kubenswrapper[4808]: I1124 18:01:03.202464 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:01:03 crc kubenswrapper[4808]: I1124 18:01:03.202805 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:01:04 crc kubenswrapper[4808]: I1124 18:01:04.087514 4808 generic.go:334] "Generic (PLEG): container finished" podID="f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e" containerID="3d4f141ef589b731e5cb42b570b8bee6520343882b3d89a7221f8b54779f2079" exitCode=0 Nov 24 18:01:04 crc kubenswrapper[4808]: I1124 18:01:04.087730 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400121-9z8sf" event={"ID":"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e","Type":"ContainerDied","Data":"3d4f141ef589b731e5cb42b570b8bee6520343882b3d89a7221f8b54779f2079"} Nov 24 18:01:04 crc kubenswrapper[4808]: I1124 18:01:04.093982 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lm2cj" event={"ID":"cc2adcf3-c591-4319-80b4-737e5aa6d117","Type":"ContainerStarted","Data":"b7d316c35748428007f920016c6d4bd8f39f4a41d94e9c45dc4884cd456f45d4"} Nov 24 18:01:04 crc kubenswrapper[4808]: I1124 18:01:04.149339 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lm2cj" podStartSLOduration=2.683841231 podStartE2EDuration="5.149308705s" podCreationTimestamp="2025-11-24 18:00:59 +0000 UTC" firstStartedPulling="2025-11-24 18:01:01.038149009 +0000 UTC m=+2053.635816851" lastFinishedPulling="2025-11-24 18:01:03.503616523 +0000 UTC m=+2056.101284325" observedRunningTime="2025-11-24 18:01:04.149278095 +0000 UTC m=+2056.746945937" watchObservedRunningTime="2025-11-24 18:01:04.149308705 +0000 UTC m=+2056.746976517" Nov 24 18:01:04 crc kubenswrapper[4808]: I1124 18:01:04.248247 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n7vbp" podUID="47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" containerName="registry-server" probeResult="failure" output=< Nov 24 18:01:04 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Nov 24 18:01:04 crc kubenswrapper[4808]: > Nov 24 18:01:05 crc kubenswrapper[4808]: I1124 18:01:05.429711 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:05 crc kubenswrapper[4808]: I1124 18:01:05.579672 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-config-data\") pod \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " Nov 24 18:01:05 crc kubenswrapper[4808]: I1124 18:01:05.579939 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjh5g\" (UniqueName: \"kubernetes.io/projected/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-kube-api-access-mjh5g\") pod \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " Nov 24 18:01:05 crc kubenswrapper[4808]: I1124 18:01:05.580520 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-combined-ca-bundle\") pod \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " Nov 24 18:01:05 crc kubenswrapper[4808]: I1124 18:01:05.580607 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-fernet-keys\") pod \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\" (UID: \"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e\") " Nov 24 18:01:05 crc kubenswrapper[4808]: I1124 18:01:05.585857 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-kube-api-access-mjh5g" (OuterVolumeSpecName: "kube-api-access-mjh5g") pod "f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e" (UID: "f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e"). InnerVolumeSpecName "kube-api-access-mjh5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:01:05 crc kubenswrapper[4808]: I1124 18:01:05.589096 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e" (UID: "f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:01:05 crc kubenswrapper[4808]: I1124 18:01:05.608888 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e" (UID: "f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:01:05 crc kubenswrapper[4808]: I1124 18:01:05.637272 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-config-data" (OuterVolumeSpecName: "config-data") pod "f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e" (UID: "f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:01:05 crc kubenswrapper[4808]: I1124 18:01:05.683470 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:05 crc kubenswrapper[4808]: I1124 18:01:05.683892 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjh5g\" (UniqueName: \"kubernetes.io/projected/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-kube-api-access-mjh5g\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:05 crc kubenswrapper[4808]: I1124 18:01:05.683904 4808 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:05 crc kubenswrapper[4808]: I1124 18:01:05.683916 4808 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:06 crc kubenswrapper[4808]: I1124 18:01:06.113928 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400121-9z8sf" event={"ID":"f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e","Type":"ContainerDied","Data":"af4400ffa0bf4591d6ffbba3d634eb56c688ff6296e5276ffb00b68e1e494f88"} Nov 24 18:01:06 crc kubenswrapper[4808]: I1124 18:01:06.113968 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af4400ffa0bf4591d6ffbba3d634eb56c688ff6296e5276ffb00b68e1e494f88" Nov 24 18:01:06 crc kubenswrapper[4808]: I1124 18:01:06.114052 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400121-9z8sf" Nov 24 18:01:06 crc kubenswrapper[4808]: I1124 18:01:06.522326 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:01:06 crc kubenswrapper[4808]: I1124 18:01:06.522380 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:01:09 crc kubenswrapper[4808]: I1124 18:01:09.976054 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:01:09 crc kubenswrapper[4808]: I1124 18:01:09.976576 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:01:10 crc kubenswrapper[4808]: I1124 18:01:10.027723 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:01:10 crc kubenswrapper[4808]: I1124 18:01:10.197008 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:01:10 crc kubenswrapper[4808]: I1124 18:01:10.267390 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lm2cj"] Nov 24 18:01:12 crc kubenswrapper[4808]: I1124 18:01:12.160325 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lm2cj" podUID="cc2adcf3-c591-4319-80b4-737e5aa6d117" containerName="registry-server" containerID="cri-o://b7d316c35748428007f920016c6d4bd8f39f4a41d94e9c45dc4884cd456f45d4" gracePeriod=2 Nov 24 18:01:12 crc kubenswrapper[4808]: I1124 18:01:12.627196 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:01:12 crc kubenswrapper[4808]: I1124 18:01:12.716974 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8hlt\" (UniqueName: \"kubernetes.io/projected/cc2adcf3-c591-4319-80b4-737e5aa6d117-kube-api-access-p8hlt\") pod \"cc2adcf3-c591-4319-80b4-737e5aa6d117\" (UID: \"cc2adcf3-c591-4319-80b4-737e5aa6d117\") " Nov 24 18:01:12 crc kubenswrapper[4808]: I1124 18:01:12.717111 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc2adcf3-c591-4319-80b4-737e5aa6d117-utilities\") pod \"cc2adcf3-c591-4319-80b4-737e5aa6d117\" (UID: \"cc2adcf3-c591-4319-80b4-737e5aa6d117\") " Nov 24 18:01:12 crc kubenswrapper[4808]: I1124 18:01:12.717512 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc2adcf3-c591-4319-80b4-737e5aa6d117-catalog-content\") pod \"cc2adcf3-c591-4319-80b4-737e5aa6d117\" (UID: \"cc2adcf3-c591-4319-80b4-737e5aa6d117\") " Nov 24 18:01:12 crc kubenswrapper[4808]: I1124 18:01:12.718128 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc2adcf3-c591-4319-80b4-737e5aa6d117-utilities" (OuterVolumeSpecName: "utilities") pod "cc2adcf3-c591-4319-80b4-737e5aa6d117" (UID: "cc2adcf3-c591-4319-80b4-737e5aa6d117"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:01:12 crc kubenswrapper[4808]: I1124 18:01:12.723313 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc2adcf3-c591-4319-80b4-737e5aa6d117-kube-api-access-p8hlt" (OuterVolumeSpecName: "kube-api-access-p8hlt") pod "cc2adcf3-c591-4319-80b4-737e5aa6d117" (UID: "cc2adcf3-c591-4319-80b4-737e5aa6d117"). InnerVolumeSpecName "kube-api-access-p8hlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:01:12 crc kubenswrapper[4808]: I1124 18:01:12.768592 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc2adcf3-c591-4319-80b4-737e5aa6d117-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc2adcf3-c591-4319-80b4-737e5aa6d117" (UID: "cc2adcf3-c591-4319-80b4-737e5aa6d117"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:01:12 crc kubenswrapper[4808]: I1124 18:01:12.819374 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8hlt\" (UniqueName: \"kubernetes.io/projected/cc2adcf3-c591-4319-80b4-737e5aa6d117-kube-api-access-p8hlt\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:12 crc kubenswrapper[4808]: I1124 18:01:12.819405 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc2adcf3-c591-4319-80b4-737e5aa6d117-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:12 crc kubenswrapper[4808]: I1124 18:01:12.819416 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc2adcf3-c591-4319-80b4-737e5aa6d117-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:13 crc kubenswrapper[4808]: I1124 18:01:13.170694 4808 generic.go:334] "Generic (PLEG): container finished" podID="cc2adcf3-c591-4319-80b4-737e5aa6d117" containerID="b7d316c35748428007f920016c6d4bd8f39f4a41d94e9c45dc4884cd456f45d4" exitCode=0 Nov 24 18:01:13 crc kubenswrapper[4808]: I1124 18:01:13.170756 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lm2cj" event={"ID":"cc2adcf3-c591-4319-80b4-737e5aa6d117","Type":"ContainerDied","Data":"b7d316c35748428007f920016c6d4bd8f39f4a41d94e9c45dc4884cd456f45d4"} Nov 24 18:01:13 crc kubenswrapper[4808]: I1124 18:01:13.170836 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lm2cj" event={"ID":"cc2adcf3-c591-4319-80b4-737e5aa6d117","Type":"ContainerDied","Data":"9acdacbf43bca14fead7907019a0180ac3061fac52f457410196835d2f70b897"} Nov 24 18:01:13 crc kubenswrapper[4808]: I1124 18:01:13.170856 4808 scope.go:117] "RemoveContainer" containerID="b7d316c35748428007f920016c6d4bd8f39f4a41d94e9c45dc4884cd456f45d4" Nov 24 18:01:13 crc kubenswrapper[4808]: I1124 18:01:13.170780 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lm2cj" Nov 24 18:01:13 crc kubenswrapper[4808]: I1124 18:01:13.190039 4808 scope.go:117] "RemoveContainer" containerID="43e875974de5b9356acd3c791a09c7ea9e40d4356903ed15b98d894163684036" Nov 24 18:01:13 crc kubenswrapper[4808]: I1124 18:01:13.209171 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lm2cj"] Nov 24 18:01:13 crc kubenswrapper[4808]: I1124 18:01:13.219638 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lm2cj"] Nov 24 18:01:13 crc kubenswrapper[4808]: I1124 18:01:13.240456 4808 scope.go:117] "RemoveContainer" containerID="f455ba8c1cc08576628feec2c770f6e2736f851cc891c7183a758d63e4ce32d1" Nov 24 18:01:13 crc kubenswrapper[4808]: I1124 18:01:13.270547 4808 scope.go:117] "RemoveContainer" containerID="b7d316c35748428007f920016c6d4bd8f39f4a41d94e9c45dc4884cd456f45d4" Nov 24 18:01:13 crc kubenswrapper[4808]: E1124 18:01:13.271672 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7d316c35748428007f920016c6d4bd8f39f4a41d94e9c45dc4884cd456f45d4\": container with ID starting with b7d316c35748428007f920016c6d4bd8f39f4a41d94e9c45dc4884cd456f45d4 not found: ID does not exist" containerID="b7d316c35748428007f920016c6d4bd8f39f4a41d94e9c45dc4884cd456f45d4" Nov 24 18:01:13 crc kubenswrapper[4808]: I1124 18:01:13.271720 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7d316c35748428007f920016c6d4bd8f39f4a41d94e9c45dc4884cd456f45d4"} err="failed to get container status \"b7d316c35748428007f920016c6d4bd8f39f4a41d94e9c45dc4884cd456f45d4\": rpc error: code = NotFound desc = could not find container \"b7d316c35748428007f920016c6d4bd8f39f4a41d94e9c45dc4884cd456f45d4\": container with ID starting with b7d316c35748428007f920016c6d4bd8f39f4a41d94e9c45dc4884cd456f45d4 not found: ID does not exist" Nov 24 18:01:13 crc kubenswrapper[4808]: I1124 18:01:13.271752 4808 scope.go:117] "RemoveContainer" containerID="43e875974de5b9356acd3c791a09c7ea9e40d4356903ed15b98d894163684036" Nov 24 18:01:13 crc kubenswrapper[4808]: E1124 18:01:13.272194 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43e875974de5b9356acd3c791a09c7ea9e40d4356903ed15b98d894163684036\": container with ID starting with 43e875974de5b9356acd3c791a09c7ea9e40d4356903ed15b98d894163684036 not found: ID does not exist" containerID="43e875974de5b9356acd3c791a09c7ea9e40d4356903ed15b98d894163684036" Nov 24 18:01:13 crc kubenswrapper[4808]: I1124 18:01:13.272225 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43e875974de5b9356acd3c791a09c7ea9e40d4356903ed15b98d894163684036"} err="failed to get container status \"43e875974de5b9356acd3c791a09c7ea9e40d4356903ed15b98d894163684036\": rpc error: code = NotFound desc = could not find container \"43e875974de5b9356acd3c791a09c7ea9e40d4356903ed15b98d894163684036\": container with ID starting with 43e875974de5b9356acd3c791a09c7ea9e40d4356903ed15b98d894163684036 not found: ID does not exist" Nov 24 18:01:13 crc kubenswrapper[4808]: I1124 18:01:13.272249 4808 scope.go:117] "RemoveContainer" containerID="f455ba8c1cc08576628feec2c770f6e2736f851cc891c7183a758d63e4ce32d1" Nov 24 18:01:13 crc kubenswrapper[4808]: E1124 18:01:13.272960 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f455ba8c1cc08576628feec2c770f6e2736f851cc891c7183a758d63e4ce32d1\": container with ID starting with f455ba8c1cc08576628feec2c770f6e2736f851cc891c7183a758d63e4ce32d1 not found: ID does not exist" containerID="f455ba8c1cc08576628feec2c770f6e2736f851cc891c7183a758d63e4ce32d1" Nov 24 18:01:13 crc kubenswrapper[4808]: I1124 18:01:13.272993 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f455ba8c1cc08576628feec2c770f6e2736f851cc891c7183a758d63e4ce32d1"} err="failed to get container status \"f455ba8c1cc08576628feec2c770f6e2736f851cc891c7183a758d63e4ce32d1\": rpc error: code = NotFound desc = could not find container \"f455ba8c1cc08576628feec2c770f6e2736f851cc891c7183a758d63e4ce32d1\": container with ID starting with f455ba8c1cc08576628feec2c770f6e2736f851cc891c7183a758d63e4ce32d1 not found: ID does not exist" Nov 24 18:01:14 crc kubenswrapper[4808]: I1124 18:01:14.263717 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n7vbp" podUID="47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" containerName="registry-server" probeResult="failure" output=< Nov 24 18:01:14 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Nov 24 18:01:14 crc kubenswrapper[4808]: > Nov 24 18:01:14 crc kubenswrapper[4808]: I1124 18:01:14.360640 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc2adcf3-c591-4319-80b4-737e5aa6d117" path="/var/lib/kubelet/pods/cc2adcf3-c591-4319-80b4-737e5aa6d117/volumes" Nov 24 18:01:23 crc kubenswrapper[4808]: I1124 18:01:23.246706 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:01:23 crc kubenswrapper[4808]: I1124 18:01:23.292035 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:01:24 crc kubenswrapper[4808]: I1124 18:01:24.025565 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n7vbp"] Nov 24 18:01:25 crc kubenswrapper[4808]: I1124 18:01:25.273370 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n7vbp" podUID="47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" containerName="registry-server" containerID="cri-o://1257e3df9f7dc2f362757f75ac9e7c7b38ba6dfd70510b308bfa4cca23493d32" gracePeriod=2 Nov 24 18:01:25 crc kubenswrapper[4808]: I1124 18:01:25.712716 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:01:25 crc kubenswrapper[4808]: I1124 18:01:25.847572 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vs84r\" (UniqueName: \"kubernetes.io/projected/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-kube-api-access-vs84r\") pod \"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5\" (UID: \"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5\") " Nov 24 18:01:25 crc kubenswrapper[4808]: I1124 18:01:25.847758 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-catalog-content\") pod \"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5\" (UID: \"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5\") " Nov 24 18:01:25 crc kubenswrapper[4808]: I1124 18:01:25.847914 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-utilities\") pod \"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5\" (UID: \"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5\") " Nov 24 18:01:25 crc kubenswrapper[4808]: I1124 18:01:25.848463 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-utilities" (OuterVolumeSpecName: "utilities") pod "47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" (UID: "47c4785a-0fc4-40c9-b9bd-45fdcd918fb5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:01:25 crc kubenswrapper[4808]: I1124 18:01:25.853415 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-kube-api-access-vs84r" (OuterVolumeSpecName: "kube-api-access-vs84r") pod "47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" (UID: "47c4785a-0fc4-40c9-b9bd-45fdcd918fb5"). InnerVolumeSpecName "kube-api-access-vs84r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:01:25 crc kubenswrapper[4808]: I1124 18:01:25.925181 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" (UID: "47c4785a-0fc4-40c9-b9bd-45fdcd918fb5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:01:25 crc kubenswrapper[4808]: I1124 18:01:25.950732 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:25 crc kubenswrapper[4808]: I1124 18:01:25.950971 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vs84r\" (UniqueName: \"kubernetes.io/projected/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-kube-api-access-vs84r\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:25 crc kubenswrapper[4808]: I1124 18:01:25.951167 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:26 crc kubenswrapper[4808]: I1124 18:01:26.286182 4808 generic.go:334] "Generic (PLEG): container finished" podID="47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" containerID="1257e3df9f7dc2f362757f75ac9e7c7b38ba6dfd70510b308bfa4cca23493d32" exitCode=0 Nov 24 18:01:26 crc kubenswrapper[4808]: I1124 18:01:26.286264 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7vbp" event={"ID":"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5","Type":"ContainerDied","Data":"1257e3df9f7dc2f362757f75ac9e7c7b38ba6dfd70510b308bfa4cca23493d32"} Nov 24 18:01:26 crc kubenswrapper[4808]: I1124 18:01:26.286273 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n7vbp" Nov 24 18:01:26 crc kubenswrapper[4808]: I1124 18:01:26.286327 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7vbp" event={"ID":"47c4785a-0fc4-40c9-b9bd-45fdcd918fb5","Type":"ContainerDied","Data":"d40a4d838348c7ef643881da1ba5eddd57d99f931bba474833c1352f901d1923"} Nov 24 18:01:26 crc kubenswrapper[4808]: I1124 18:01:26.286361 4808 scope.go:117] "RemoveContainer" containerID="1257e3df9f7dc2f362757f75ac9e7c7b38ba6dfd70510b308bfa4cca23493d32" Nov 24 18:01:26 crc kubenswrapper[4808]: I1124 18:01:26.320521 4808 scope.go:117] "RemoveContainer" containerID="0284776c8a628f219687045d70ec892dcea3ef840d71b8d59931376d99674306" Nov 24 18:01:26 crc kubenswrapper[4808]: I1124 18:01:26.348791 4808 scope.go:117] "RemoveContainer" containerID="5e6e66cad2aa309536c44c69b95e67bab23173965451f11d09b574d046457fc5" Nov 24 18:01:26 crc kubenswrapper[4808]: I1124 18:01:26.367312 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n7vbp"] Nov 24 18:01:26 crc kubenswrapper[4808]: I1124 18:01:26.370062 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n7vbp"] Nov 24 18:01:26 crc kubenswrapper[4808]: I1124 18:01:26.417687 4808 scope.go:117] "RemoveContainer" containerID="1257e3df9f7dc2f362757f75ac9e7c7b38ba6dfd70510b308bfa4cca23493d32" Nov 24 18:01:26 crc kubenswrapper[4808]: E1124 18:01:26.418070 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1257e3df9f7dc2f362757f75ac9e7c7b38ba6dfd70510b308bfa4cca23493d32\": container with ID starting with 1257e3df9f7dc2f362757f75ac9e7c7b38ba6dfd70510b308bfa4cca23493d32 not found: ID does not exist" containerID="1257e3df9f7dc2f362757f75ac9e7c7b38ba6dfd70510b308bfa4cca23493d32" Nov 24 18:01:26 crc kubenswrapper[4808]: I1124 18:01:26.418113 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1257e3df9f7dc2f362757f75ac9e7c7b38ba6dfd70510b308bfa4cca23493d32"} err="failed to get container status \"1257e3df9f7dc2f362757f75ac9e7c7b38ba6dfd70510b308bfa4cca23493d32\": rpc error: code = NotFound desc = could not find container \"1257e3df9f7dc2f362757f75ac9e7c7b38ba6dfd70510b308bfa4cca23493d32\": container with ID starting with 1257e3df9f7dc2f362757f75ac9e7c7b38ba6dfd70510b308bfa4cca23493d32 not found: ID does not exist" Nov 24 18:01:26 crc kubenswrapper[4808]: I1124 18:01:26.418141 4808 scope.go:117] "RemoveContainer" containerID="0284776c8a628f219687045d70ec892dcea3ef840d71b8d59931376d99674306" Nov 24 18:01:26 crc kubenswrapper[4808]: E1124 18:01:26.418543 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0284776c8a628f219687045d70ec892dcea3ef840d71b8d59931376d99674306\": container with ID starting with 0284776c8a628f219687045d70ec892dcea3ef840d71b8d59931376d99674306 not found: ID does not exist" containerID="0284776c8a628f219687045d70ec892dcea3ef840d71b8d59931376d99674306" Nov 24 18:01:26 crc kubenswrapper[4808]: I1124 18:01:26.418572 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0284776c8a628f219687045d70ec892dcea3ef840d71b8d59931376d99674306"} err="failed to get container status \"0284776c8a628f219687045d70ec892dcea3ef840d71b8d59931376d99674306\": rpc error: code = NotFound desc = could not find container \"0284776c8a628f219687045d70ec892dcea3ef840d71b8d59931376d99674306\": container with ID starting with 0284776c8a628f219687045d70ec892dcea3ef840d71b8d59931376d99674306 not found: ID does not exist" Nov 24 18:01:26 crc kubenswrapper[4808]: I1124 18:01:26.418597 4808 scope.go:117] "RemoveContainer" containerID="5e6e66cad2aa309536c44c69b95e67bab23173965451f11d09b574d046457fc5" Nov 24 18:01:26 crc kubenswrapper[4808]: E1124 18:01:26.418894 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e6e66cad2aa309536c44c69b95e67bab23173965451f11d09b574d046457fc5\": container with ID starting with 5e6e66cad2aa309536c44c69b95e67bab23173965451f11d09b574d046457fc5 not found: ID does not exist" containerID="5e6e66cad2aa309536c44c69b95e67bab23173965451f11d09b574d046457fc5" Nov 24 18:01:26 crc kubenswrapper[4808]: I1124 18:01:26.418912 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e6e66cad2aa309536c44c69b95e67bab23173965451f11d09b574d046457fc5"} err="failed to get container status \"5e6e66cad2aa309536c44c69b95e67bab23173965451f11d09b574d046457fc5\": rpc error: code = NotFound desc = could not find container \"5e6e66cad2aa309536c44c69b95e67bab23173965451f11d09b574d046457fc5\": container with ID starting with 5e6e66cad2aa309536c44c69b95e67bab23173965451f11d09b574d046457fc5 not found: ID does not exist" Nov 24 18:01:28 crc kubenswrapper[4808]: I1124 18:01:28.358275 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" path="/var/lib/kubelet/pods/47c4785a-0fc4-40c9-b9bd-45fdcd918fb5/volumes" Nov 24 18:01:36 crc kubenswrapper[4808]: I1124 18:01:36.522927 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:01:36 crc kubenswrapper[4808]: I1124 18:01:36.524687 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.020853 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mh59l"] Nov 24 18:01:40 crc kubenswrapper[4808]: E1124 18:01:40.021609 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" containerName="extract-utilities" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.021624 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" containerName="extract-utilities" Nov 24 18:01:40 crc kubenswrapper[4808]: E1124 18:01:40.021637 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc2adcf3-c591-4319-80b4-737e5aa6d117" containerName="registry-server" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.021643 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc2adcf3-c591-4319-80b4-737e5aa6d117" containerName="registry-server" Nov 24 18:01:40 crc kubenswrapper[4808]: E1124 18:01:40.021657 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" containerName="registry-server" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.021664 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" containerName="registry-server" Nov 24 18:01:40 crc kubenswrapper[4808]: E1124 18:01:40.021675 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc2adcf3-c591-4319-80b4-737e5aa6d117" containerName="extract-utilities" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.021682 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc2adcf3-c591-4319-80b4-737e5aa6d117" containerName="extract-utilities" Nov 24 18:01:40 crc kubenswrapper[4808]: E1124 18:01:40.021691 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" containerName="extract-content" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.021697 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" containerName="extract-content" Nov 24 18:01:40 crc kubenswrapper[4808]: E1124 18:01:40.021722 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e" containerName="keystone-cron" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.021728 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e" containerName="keystone-cron" Nov 24 18:01:40 crc kubenswrapper[4808]: E1124 18:01:40.021737 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc2adcf3-c591-4319-80b4-737e5aa6d117" containerName="extract-content" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.021744 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc2adcf3-c591-4319-80b4-737e5aa6d117" containerName="extract-content" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.021949 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="47c4785a-0fc4-40c9-b9bd-45fdcd918fb5" containerName="registry-server" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.021969 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e" containerName="keystone-cron" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.021982 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc2adcf3-c591-4319-80b4-737e5aa6d117" containerName="registry-server" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.023818 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.027448 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f388fde-3747-4229-987e-bd6036a912c4-utilities\") pod \"certified-operators-mh59l\" (UID: \"3f388fde-3747-4229-987e-bd6036a912c4\") " pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.027567 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xk9m\" (UniqueName: \"kubernetes.io/projected/3f388fde-3747-4229-987e-bd6036a912c4-kube-api-access-7xk9m\") pod \"certified-operators-mh59l\" (UID: \"3f388fde-3747-4229-987e-bd6036a912c4\") " pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.027617 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f388fde-3747-4229-987e-bd6036a912c4-catalog-content\") pod \"certified-operators-mh59l\" (UID: \"3f388fde-3747-4229-987e-bd6036a912c4\") " pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.037197 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mh59l"] Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.129054 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xk9m\" (UniqueName: \"kubernetes.io/projected/3f388fde-3747-4229-987e-bd6036a912c4-kube-api-access-7xk9m\") pod \"certified-operators-mh59l\" (UID: \"3f388fde-3747-4229-987e-bd6036a912c4\") " pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.129122 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f388fde-3747-4229-987e-bd6036a912c4-catalog-content\") pod \"certified-operators-mh59l\" (UID: \"3f388fde-3747-4229-987e-bd6036a912c4\") " pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.129169 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f388fde-3747-4229-987e-bd6036a912c4-utilities\") pod \"certified-operators-mh59l\" (UID: \"3f388fde-3747-4229-987e-bd6036a912c4\") " pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.129804 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f388fde-3747-4229-987e-bd6036a912c4-catalog-content\") pod \"certified-operators-mh59l\" (UID: \"3f388fde-3747-4229-987e-bd6036a912c4\") " pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.129944 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f388fde-3747-4229-987e-bd6036a912c4-utilities\") pod \"certified-operators-mh59l\" (UID: \"3f388fde-3747-4229-987e-bd6036a912c4\") " pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.166826 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xk9m\" (UniqueName: \"kubernetes.io/projected/3f388fde-3747-4229-987e-bd6036a912c4-kube-api-access-7xk9m\") pod \"certified-operators-mh59l\" (UID: \"3f388fde-3747-4229-987e-bd6036a912c4\") " pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.353114 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:40 crc kubenswrapper[4808]: I1124 18:01:40.906858 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mh59l"] Nov 24 18:01:41 crc kubenswrapper[4808]: I1124 18:01:41.425201 4808 generic.go:334] "Generic (PLEG): container finished" podID="3f388fde-3747-4229-987e-bd6036a912c4" containerID="0366a9ae9b66ec400d67422c5b371bf4bf2c412563ee9435dd5981a706b3b448" exitCode=0 Nov 24 18:01:41 crc kubenswrapper[4808]: I1124 18:01:41.425290 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mh59l" event={"ID":"3f388fde-3747-4229-987e-bd6036a912c4","Type":"ContainerDied","Data":"0366a9ae9b66ec400d67422c5b371bf4bf2c412563ee9435dd5981a706b3b448"} Nov 24 18:01:41 crc kubenswrapper[4808]: I1124 18:01:41.425585 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mh59l" event={"ID":"3f388fde-3747-4229-987e-bd6036a912c4","Type":"ContainerStarted","Data":"cfafcf47ad7f426769aea97117c7b42c86b6007d853d5a158d61815ee66c045b"} Nov 24 18:01:43 crc kubenswrapper[4808]: I1124 18:01:43.443705 4808 generic.go:334] "Generic (PLEG): container finished" podID="3f388fde-3747-4229-987e-bd6036a912c4" containerID="dc0f7f13dff758300a9b379776965387a421bedce54383559a8f3b9bdfa8bb33" exitCode=0 Nov 24 18:01:43 crc kubenswrapper[4808]: I1124 18:01:43.443762 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mh59l" event={"ID":"3f388fde-3747-4229-987e-bd6036a912c4","Type":"ContainerDied","Data":"dc0f7f13dff758300a9b379776965387a421bedce54383559a8f3b9bdfa8bb33"} Nov 24 18:01:44 crc kubenswrapper[4808]: I1124 18:01:44.453258 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mh59l" event={"ID":"3f388fde-3747-4229-987e-bd6036a912c4","Type":"ContainerStarted","Data":"410e893969e0b2179baed5f3bed651ae5c4f736acfea76332eb6a5b1cab93efc"} Nov 24 18:01:44 crc kubenswrapper[4808]: I1124 18:01:44.480999 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mh59l" podStartSLOduration=2.990531511 podStartE2EDuration="5.480982208s" podCreationTimestamp="2025-11-24 18:01:39 +0000 UTC" firstStartedPulling="2025-11-24 18:01:41.426897603 +0000 UTC m=+2094.024565405" lastFinishedPulling="2025-11-24 18:01:43.9173483 +0000 UTC m=+2096.515016102" observedRunningTime="2025-11-24 18:01:44.480917986 +0000 UTC m=+2097.078585788" watchObservedRunningTime="2025-11-24 18:01:44.480982208 +0000 UTC m=+2097.078650010" Nov 24 18:01:50 crc kubenswrapper[4808]: I1124 18:01:50.379649 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:50 crc kubenswrapper[4808]: I1124 18:01:50.380123 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:50 crc kubenswrapper[4808]: I1124 18:01:50.410641 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:50 crc kubenswrapper[4808]: I1124 18:01:50.551009 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:50 crc kubenswrapper[4808]: I1124 18:01:50.645069 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mh59l"] Nov 24 18:01:52 crc kubenswrapper[4808]: I1124 18:01:52.521436 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mh59l" podUID="3f388fde-3747-4229-987e-bd6036a912c4" containerName="registry-server" containerID="cri-o://410e893969e0b2179baed5f3bed651ae5c4f736acfea76332eb6a5b1cab93efc" gracePeriod=2 Nov 24 18:01:52 crc kubenswrapper[4808]: I1124 18:01:52.932786 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.077113 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f388fde-3747-4229-987e-bd6036a912c4-utilities\") pod \"3f388fde-3747-4229-987e-bd6036a912c4\" (UID: \"3f388fde-3747-4229-987e-bd6036a912c4\") " Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.077259 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f388fde-3747-4229-987e-bd6036a912c4-catalog-content\") pod \"3f388fde-3747-4229-987e-bd6036a912c4\" (UID: \"3f388fde-3747-4229-987e-bd6036a912c4\") " Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.077309 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xk9m\" (UniqueName: \"kubernetes.io/projected/3f388fde-3747-4229-987e-bd6036a912c4-kube-api-access-7xk9m\") pod \"3f388fde-3747-4229-987e-bd6036a912c4\" (UID: \"3f388fde-3747-4229-987e-bd6036a912c4\") " Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.079769 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f388fde-3747-4229-987e-bd6036a912c4-utilities" (OuterVolumeSpecName: "utilities") pod "3f388fde-3747-4229-987e-bd6036a912c4" (UID: "3f388fde-3747-4229-987e-bd6036a912c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.084138 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f388fde-3747-4229-987e-bd6036a912c4-kube-api-access-7xk9m" (OuterVolumeSpecName: "kube-api-access-7xk9m") pod "3f388fde-3747-4229-987e-bd6036a912c4" (UID: "3f388fde-3747-4229-987e-bd6036a912c4"). InnerVolumeSpecName "kube-api-access-7xk9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.180144 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f388fde-3747-4229-987e-bd6036a912c4-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.180201 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xk9m\" (UniqueName: \"kubernetes.io/projected/3f388fde-3747-4229-987e-bd6036a912c4-kube-api-access-7xk9m\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.315874 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f388fde-3747-4229-987e-bd6036a912c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f388fde-3747-4229-987e-bd6036a912c4" (UID: "3f388fde-3747-4229-987e-bd6036a912c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.384000 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f388fde-3747-4229-987e-bd6036a912c4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.534981 4808 generic.go:334] "Generic (PLEG): container finished" podID="3f388fde-3747-4229-987e-bd6036a912c4" containerID="410e893969e0b2179baed5f3bed651ae5c4f736acfea76332eb6a5b1cab93efc" exitCode=0 Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.535058 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mh59l" event={"ID":"3f388fde-3747-4229-987e-bd6036a912c4","Type":"ContainerDied","Data":"410e893969e0b2179baed5f3bed651ae5c4f736acfea76332eb6a5b1cab93efc"} Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.535088 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mh59l" event={"ID":"3f388fde-3747-4229-987e-bd6036a912c4","Type":"ContainerDied","Data":"cfafcf47ad7f426769aea97117c7b42c86b6007d853d5a158d61815ee66c045b"} Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.535107 4808 scope.go:117] "RemoveContainer" containerID="410e893969e0b2179baed5f3bed651ae5c4f736acfea76332eb6a5b1cab93efc" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.535137 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mh59l" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.553590 4808 scope.go:117] "RemoveContainer" containerID="dc0f7f13dff758300a9b379776965387a421bedce54383559a8f3b9bdfa8bb33" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.572497 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mh59l"] Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.588981 4808 scope.go:117] "RemoveContainer" containerID="0366a9ae9b66ec400d67422c5b371bf4bf2c412563ee9435dd5981a706b3b448" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.597570 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mh59l"] Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.629931 4808 scope.go:117] "RemoveContainer" containerID="410e893969e0b2179baed5f3bed651ae5c4f736acfea76332eb6a5b1cab93efc" Nov 24 18:01:53 crc kubenswrapper[4808]: E1124 18:01:53.630411 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"410e893969e0b2179baed5f3bed651ae5c4f736acfea76332eb6a5b1cab93efc\": container with ID starting with 410e893969e0b2179baed5f3bed651ae5c4f736acfea76332eb6a5b1cab93efc not found: ID does not exist" containerID="410e893969e0b2179baed5f3bed651ae5c4f736acfea76332eb6a5b1cab93efc" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.630455 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"410e893969e0b2179baed5f3bed651ae5c4f736acfea76332eb6a5b1cab93efc"} err="failed to get container status \"410e893969e0b2179baed5f3bed651ae5c4f736acfea76332eb6a5b1cab93efc\": rpc error: code = NotFound desc = could not find container \"410e893969e0b2179baed5f3bed651ae5c4f736acfea76332eb6a5b1cab93efc\": container with ID starting with 410e893969e0b2179baed5f3bed651ae5c4f736acfea76332eb6a5b1cab93efc not found: ID does not exist" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.630483 4808 scope.go:117] "RemoveContainer" containerID="dc0f7f13dff758300a9b379776965387a421bedce54383559a8f3b9bdfa8bb33" Nov 24 18:01:53 crc kubenswrapper[4808]: E1124 18:01:53.630746 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc0f7f13dff758300a9b379776965387a421bedce54383559a8f3b9bdfa8bb33\": container with ID starting with dc0f7f13dff758300a9b379776965387a421bedce54383559a8f3b9bdfa8bb33 not found: ID does not exist" containerID="dc0f7f13dff758300a9b379776965387a421bedce54383559a8f3b9bdfa8bb33" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.630774 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0f7f13dff758300a9b379776965387a421bedce54383559a8f3b9bdfa8bb33"} err="failed to get container status \"dc0f7f13dff758300a9b379776965387a421bedce54383559a8f3b9bdfa8bb33\": rpc error: code = NotFound desc = could not find container \"dc0f7f13dff758300a9b379776965387a421bedce54383559a8f3b9bdfa8bb33\": container with ID starting with dc0f7f13dff758300a9b379776965387a421bedce54383559a8f3b9bdfa8bb33 not found: ID does not exist" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.630790 4808 scope.go:117] "RemoveContainer" containerID="0366a9ae9b66ec400d67422c5b371bf4bf2c412563ee9435dd5981a706b3b448" Nov 24 18:01:53 crc kubenswrapper[4808]: E1124 18:01:53.631062 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0366a9ae9b66ec400d67422c5b371bf4bf2c412563ee9435dd5981a706b3b448\": container with ID starting with 0366a9ae9b66ec400d67422c5b371bf4bf2c412563ee9435dd5981a706b3b448 not found: ID does not exist" containerID="0366a9ae9b66ec400d67422c5b371bf4bf2c412563ee9435dd5981a706b3b448" Nov 24 18:01:53 crc kubenswrapper[4808]: I1124 18:01:53.631090 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0366a9ae9b66ec400d67422c5b371bf4bf2c412563ee9435dd5981a706b3b448"} err="failed to get container status \"0366a9ae9b66ec400d67422c5b371bf4bf2c412563ee9435dd5981a706b3b448\": rpc error: code = NotFound desc = could not find container \"0366a9ae9b66ec400d67422c5b371bf4bf2c412563ee9435dd5981a706b3b448\": container with ID starting with 0366a9ae9b66ec400d67422c5b371bf4bf2c412563ee9435dd5981a706b3b448 not found: ID does not exist" Nov 24 18:01:54 crc kubenswrapper[4808]: I1124 18:01:54.359850 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f388fde-3747-4229-987e-bd6036a912c4" path="/var/lib/kubelet/pods/3f388fde-3747-4229-987e-bd6036a912c4/volumes" Nov 24 18:02:06 crc kubenswrapper[4808]: I1124 18:02:06.523429 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:02:06 crc kubenswrapper[4808]: I1124 18:02:06.524694 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:02:06 crc kubenswrapper[4808]: I1124 18:02:06.524785 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 18:02:06 crc kubenswrapper[4808]: I1124 18:02:06.526097 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a01acbe1d0e5b11ae5eb995ee52c951450e7a150c9c344c5eacfa7fe2bda2db2"} pod="openshift-machine-config-operator/machine-config-daemon-ht87w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 18:02:06 crc kubenswrapper[4808]: I1124 18:02:06.526222 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" containerID="cri-o://a01acbe1d0e5b11ae5eb995ee52c951450e7a150c9c344c5eacfa7fe2bda2db2" gracePeriod=600 Nov 24 18:02:07 crc kubenswrapper[4808]: I1124 18:02:07.946086 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerID="a01acbe1d0e5b11ae5eb995ee52c951450e7a150c9c344c5eacfa7fe2bda2db2" exitCode=0 Nov 24 18:02:07 crc kubenswrapper[4808]: I1124 18:02:07.946147 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerDied","Data":"a01acbe1d0e5b11ae5eb995ee52c951450e7a150c9c344c5eacfa7fe2bda2db2"} Nov 24 18:02:07 crc kubenswrapper[4808]: I1124 18:02:07.946697 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f"} Nov 24 18:02:07 crc kubenswrapper[4808]: I1124 18:02:07.946719 4808 scope.go:117] "RemoveContainer" containerID="84472d9753b42cd4349b8d748296967c25570c9eef6aacda1bc1a71dfbe5500f" Nov 24 18:02:52 crc kubenswrapper[4808]: I1124 18:02:52.354707 4808 generic.go:334] "Generic (PLEG): container finished" podID="3a26fcdc-f080-4677-a308-b08584153734" containerID="cad27741ed70c1921571017076dc2e7ee1b42883f999a6896d5ac43fa374246e" exitCode=0 Nov 24 18:02:52 crc kubenswrapper[4808]: I1124 18:02:52.362821 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" event={"ID":"3a26fcdc-f080-4677-a308-b08584153734","Type":"ContainerDied","Data":"cad27741ed70c1921571017076dc2e7ee1b42883f999a6896d5ac43fa374246e"} Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.768994 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.886976 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sws4p\" (UniqueName: \"kubernetes.io/projected/3a26fcdc-f080-4677-a308-b08584153734-kube-api-access-sws4p\") pod \"3a26fcdc-f080-4677-a308-b08584153734\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.887592 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-libvirt-combined-ca-bundle\") pod \"3a26fcdc-f080-4677-a308-b08584153734\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.887659 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-inventory\") pod \"3a26fcdc-f080-4677-a308-b08584153734\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.887960 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-ssh-key\") pod \"3a26fcdc-f080-4677-a308-b08584153734\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.888122 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-libvirt-secret-0\") pod \"3a26fcdc-f080-4677-a308-b08584153734\" (UID: \"3a26fcdc-f080-4677-a308-b08584153734\") " Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.893175 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "3a26fcdc-f080-4677-a308-b08584153734" (UID: "3a26fcdc-f080-4677-a308-b08584153734"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.896513 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a26fcdc-f080-4677-a308-b08584153734-kube-api-access-sws4p" (OuterVolumeSpecName: "kube-api-access-sws4p") pod "3a26fcdc-f080-4677-a308-b08584153734" (UID: "3a26fcdc-f080-4677-a308-b08584153734"). InnerVolumeSpecName "kube-api-access-sws4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.919270 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3a26fcdc-f080-4677-a308-b08584153734" (UID: "3a26fcdc-f080-4677-a308-b08584153734"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.937503 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "3a26fcdc-f080-4677-a308-b08584153734" (UID: "3a26fcdc-f080-4677-a308-b08584153734"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.938743 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-inventory" (OuterVolumeSpecName: "inventory") pod "3a26fcdc-f080-4677-a308-b08584153734" (UID: "3a26fcdc-f080-4677-a308-b08584153734"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.990124 4808 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.990161 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sws4p\" (UniqueName: \"kubernetes.io/projected/3a26fcdc-f080-4677-a308-b08584153734-kube-api-access-sws4p\") on node \"crc\" DevicePath \"\"" Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.990178 4808 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.990190 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 18:02:53 crc kubenswrapper[4808]: I1124 18:02:53.990202 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a26fcdc-f080-4677-a308-b08584153734-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.382803 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" event={"ID":"3a26fcdc-f080-4677-a308-b08584153734","Type":"ContainerDied","Data":"55776be9c796f5071c8cc1bbc6e856dd7b96590762dac3d637952e25e8b4e3bf"} Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.382852 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.382859 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55776be9c796f5071c8cc1bbc6e856dd7b96590762dac3d637952e25e8b4e3bf" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.496067 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh"] Nov 24 18:02:54 crc kubenswrapper[4808]: E1124 18:02:54.496411 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f388fde-3747-4229-987e-bd6036a912c4" containerName="extract-content" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.496423 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f388fde-3747-4229-987e-bd6036a912c4" containerName="extract-content" Nov 24 18:02:54 crc kubenswrapper[4808]: E1124 18:02:54.496440 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a26fcdc-f080-4677-a308-b08584153734" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.496447 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a26fcdc-f080-4677-a308-b08584153734" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 18:02:54 crc kubenswrapper[4808]: E1124 18:02:54.496467 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f388fde-3747-4229-987e-bd6036a912c4" containerName="registry-server" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.496475 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f388fde-3747-4229-987e-bd6036a912c4" containerName="registry-server" Nov 24 18:02:54 crc kubenswrapper[4808]: E1124 18:02:54.496495 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f388fde-3747-4229-987e-bd6036a912c4" containerName="extract-utilities" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.496502 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f388fde-3747-4229-987e-bd6036a912c4" containerName="extract-utilities" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.496668 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f388fde-3747-4229-987e-bd6036a912c4" containerName="registry-server" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.496687 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a26fcdc-f080-4677-a308-b08584153734" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.497534 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.499460 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.499665 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.500114 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.500192 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.500494 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.500552 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.500794 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.555403 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh"] Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.601205 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.601284 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.601346 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.601410 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.601475 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.601545 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.601585 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.601605 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntw5l\" (UniqueName: \"kubernetes.io/projected/e03cdbd6-8f86-468f-a798-961c6a4920d7-kube-api-access-ntw5l\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.601643 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.702982 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.703147 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.703185 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.703213 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntw5l\" (UniqueName: \"kubernetes.io/projected/e03cdbd6-8f86-468f-a798-961c6a4920d7-kube-api-access-ntw5l\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.703263 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.703285 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.703322 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.703359 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.703414 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.705176 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.709857 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.710764 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.711448 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.711952 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.716869 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.721824 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.723363 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntw5l\" (UniqueName: \"kubernetes.io/projected/e03cdbd6-8f86-468f-a798-961c6a4920d7-kube-api-access-ntw5l\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.727074 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lfxmh\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:54 crc kubenswrapper[4808]: I1124 18:02:54.815847 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:02:55 crc kubenswrapper[4808]: I1124 18:02:55.318432 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh"] Nov 24 18:02:55 crc kubenswrapper[4808]: I1124 18:02:55.395114 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" event={"ID":"e03cdbd6-8f86-468f-a798-961c6a4920d7","Type":"ContainerStarted","Data":"31fb9b7853097eddf93e8b7cf5f3b23b22147a34441b836dae8fee7100d72049"} Nov 24 18:02:56 crc kubenswrapper[4808]: I1124 18:02:56.406085 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" event={"ID":"e03cdbd6-8f86-468f-a798-961c6a4920d7","Type":"ContainerStarted","Data":"824e7f71e1ed99ae2cb4f76fd6903bf5c887d80e0f148be270f745a9711f8b7c"} Nov 24 18:02:56 crc kubenswrapper[4808]: I1124 18:02:56.427580 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" podStartSLOduration=2.018156077 podStartE2EDuration="2.427559903s" podCreationTimestamp="2025-11-24 18:02:54 +0000 UTC" firstStartedPulling="2025-11-24 18:02:55.320836525 +0000 UTC m=+2167.918504347" lastFinishedPulling="2025-11-24 18:02:55.730240331 +0000 UTC m=+2168.327908173" observedRunningTime="2025-11-24 18:02:56.419713911 +0000 UTC m=+2169.017381763" watchObservedRunningTime="2025-11-24 18:02:56.427559903 +0000 UTC m=+2169.025227705" Nov 24 18:04:36 crc kubenswrapper[4808]: I1124 18:04:36.522957 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:04:36 crc kubenswrapper[4808]: I1124 18:04:36.524216 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:04:41 crc kubenswrapper[4808]: I1124 18:04:41.794803 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c54tv"] Nov 24 18:04:41 crc kubenswrapper[4808]: I1124 18:04:41.798364 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:41 crc kubenswrapper[4808]: I1124 18:04:41.807582 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c54tv"] Nov 24 18:04:41 crc kubenswrapper[4808]: I1124 18:04:41.943130 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9b490b3-a462-4471-b9ea-ed125f0e4de8-utilities\") pod \"redhat-marketplace-c54tv\" (UID: \"b9b490b3-a462-4471-b9ea-ed125f0e4de8\") " pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:41 crc kubenswrapper[4808]: I1124 18:04:41.943362 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blppr\" (UniqueName: \"kubernetes.io/projected/b9b490b3-a462-4471-b9ea-ed125f0e4de8-kube-api-access-blppr\") pod \"redhat-marketplace-c54tv\" (UID: \"b9b490b3-a462-4471-b9ea-ed125f0e4de8\") " pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:41 crc kubenswrapper[4808]: I1124 18:04:41.943655 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9b490b3-a462-4471-b9ea-ed125f0e4de8-catalog-content\") pod \"redhat-marketplace-c54tv\" (UID: \"b9b490b3-a462-4471-b9ea-ed125f0e4de8\") " pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:42 crc kubenswrapper[4808]: I1124 18:04:42.045239 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blppr\" (UniqueName: \"kubernetes.io/projected/b9b490b3-a462-4471-b9ea-ed125f0e4de8-kube-api-access-blppr\") pod \"redhat-marketplace-c54tv\" (UID: \"b9b490b3-a462-4471-b9ea-ed125f0e4de8\") " pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:42 crc kubenswrapper[4808]: I1124 18:04:42.045694 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9b490b3-a462-4471-b9ea-ed125f0e4de8-catalog-content\") pod \"redhat-marketplace-c54tv\" (UID: \"b9b490b3-a462-4471-b9ea-ed125f0e4de8\") " pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:42 crc kubenswrapper[4808]: I1124 18:04:42.045793 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9b490b3-a462-4471-b9ea-ed125f0e4de8-utilities\") pod \"redhat-marketplace-c54tv\" (UID: \"b9b490b3-a462-4471-b9ea-ed125f0e4de8\") " pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:42 crc kubenswrapper[4808]: I1124 18:04:42.046313 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9b490b3-a462-4471-b9ea-ed125f0e4de8-catalog-content\") pod \"redhat-marketplace-c54tv\" (UID: \"b9b490b3-a462-4471-b9ea-ed125f0e4de8\") " pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:42 crc kubenswrapper[4808]: I1124 18:04:42.046447 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9b490b3-a462-4471-b9ea-ed125f0e4de8-utilities\") pod \"redhat-marketplace-c54tv\" (UID: \"b9b490b3-a462-4471-b9ea-ed125f0e4de8\") " pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:42 crc kubenswrapper[4808]: I1124 18:04:42.071482 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blppr\" (UniqueName: \"kubernetes.io/projected/b9b490b3-a462-4471-b9ea-ed125f0e4de8-kube-api-access-blppr\") pod \"redhat-marketplace-c54tv\" (UID: \"b9b490b3-a462-4471-b9ea-ed125f0e4de8\") " pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:42 crc kubenswrapper[4808]: I1124 18:04:42.118716 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:42 crc kubenswrapper[4808]: I1124 18:04:42.579701 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c54tv"] Nov 24 18:04:43 crc kubenswrapper[4808]: I1124 18:04:43.298316 4808 generic.go:334] "Generic (PLEG): container finished" podID="b9b490b3-a462-4471-b9ea-ed125f0e4de8" containerID="1c46a2cf91be07d96b0fe7fffe31f04cd18331ec7ea378a5a84d6ebc5f921ba5" exitCode=0 Nov 24 18:04:43 crc kubenswrapper[4808]: I1124 18:04:43.298364 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c54tv" event={"ID":"b9b490b3-a462-4471-b9ea-ed125f0e4de8","Type":"ContainerDied","Data":"1c46a2cf91be07d96b0fe7fffe31f04cd18331ec7ea378a5a84d6ebc5f921ba5"} Nov 24 18:04:43 crc kubenswrapper[4808]: I1124 18:04:43.298655 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c54tv" event={"ID":"b9b490b3-a462-4471-b9ea-ed125f0e4de8","Type":"ContainerStarted","Data":"8ab900936615333ecc4c087a70a96c15db7cad1e69bf463746b4555ede1c25da"} Nov 24 18:04:43 crc kubenswrapper[4808]: I1124 18:04:43.300852 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 18:04:44 crc kubenswrapper[4808]: I1124 18:04:44.310195 4808 generic.go:334] "Generic (PLEG): container finished" podID="b9b490b3-a462-4471-b9ea-ed125f0e4de8" containerID="ce2a7a4cdaa1027d3932fd8b52c941ddcb6f864bd936a7c276c7f7686a8efd3b" exitCode=0 Nov 24 18:04:44 crc kubenswrapper[4808]: I1124 18:04:44.310260 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c54tv" event={"ID":"b9b490b3-a462-4471-b9ea-ed125f0e4de8","Type":"ContainerDied","Data":"ce2a7a4cdaa1027d3932fd8b52c941ddcb6f864bd936a7c276c7f7686a8efd3b"} Nov 24 18:04:45 crc kubenswrapper[4808]: I1124 18:04:45.345639 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c54tv" event={"ID":"b9b490b3-a462-4471-b9ea-ed125f0e4de8","Type":"ContainerStarted","Data":"84622f3612c0e8e1668729dadc088c0e107eef001151a69dd0daf8fad47031c7"} Nov 24 18:04:45 crc kubenswrapper[4808]: I1124 18:04:45.377509 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c54tv" podStartSLOduration=2.975404213 podStartE2EDuration="4.377476524s" podCreationTimestamp="2025-11-24 18:04:41 +0000 UTC" firstStartedPulling="2025-11-24 18:04:43.300547146 +0000 UTC m=+2275.898214948" lastFinishedPulling="2025-11-24 18:04:44.702619457 +0000 UTC m=+2277.300287259" observedRunningTime="2025-11-24 18:04:45.367970447 +0000 UTC m=+2277.965638269" watchObservedRunningTime="2025-11-24 18:04:45.377476524 +0000 UTC m=+2277.975144326" Nov 24 18:04:52 crc kubenswrapper[4808]: I1124 18:04:52.119519 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:52 crc kubenswrapper[4808]: I1124 18:04:52.120109 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:52 crc kubenswrapper[4808]: I1124 18:04:52.167552 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:52 crc kubenswrapper[4808]: I1124 18:04:52.499449 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:52 crc kubenswrapper[4808]: I1124 18:04:52.559211 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c54tv"] Nov 24 18:04:54 crc kubenswrapper[4808]: I1124 18:04:54.456863 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c54tv" podUID="b9b490b3-a462-4471-b9ea-ed125f0e4de8" containerName="registry-server" containerID="cri-o://84622f3612c0e8e1668729dadc088c0e107eef001151a69dd0daf8fad47031c7" gracePeriod=2 Nov 24 18:04:54 crc kubenswrapper[4808]: I1124 18:04:54.962618 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.032810 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9b490b3-a462-4471-b9ea-ed125f0e4de8-utilities\") pod \"b9b490b3-a462-4471-b9ea-ed125f0e4de8\" (UID: \"b9b490b3-a462-4471-b9ea-ed125f0e4de8\") " Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.032886 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9b490b3-a462-4471-b9ea-ed125f0e4de8-catalog-content\") pod \"b9b490b3-a462-4471-b9ea-ed125f0e4de8\" (UID: \"b9b490b3-a462-4471-b9ea-ed125f0e4de8\") " Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.032989 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blppr\" (UniqueName: \"kubernetes.io/projected/b9b490b3-a462-4471-b9ea-ed125f0e4de8-kube-api-access-blppr\") pod \"b9b490b3-a462-4471-b9ea-ed125f0e4de8\" (UID: \"b9b490b3-a462-4471-b9ea-ed125f0e4de8\") " Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.033984 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9b490b3-a462-4471-b9ea-ed125f0e4de8-utilities" (OuterVolumeSpecName: "utilities") pod "b9b490b3-a462-4471-b9ea-ed125f0e4de8" (UID: "b9b490b3-a462-4471-b9ea-ed125f0e4de8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.038193 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9b490b3-a462-4471-b9ea-ed125f0e4de8-kube-api-access-blppr" (OuterVolumeSpecName: "kube-api-access-blppr") pod "b9b490b3-a462-4471-b9ea-ed125f0e4de8" (UID: "b9b490b3-a462-4471-b9ea-ed125f0e4de8"). InnerVolumeSpecName "kube-api-access-blppr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.052139 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9b490b3-a462-4471-b9ea-ed125f0e4de8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9b490b3-a462-4471-b9ea-ed125f0e4de8" (UID: "b9b490b3-a462-4471-b9ea-ed125f0e4de8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.134976 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blppr\" (UniqueName: \"kubernetes.io/projected/b9b490b3-a462-4471-b9ea-ed125f0e4de8-kube-api-access-blppr\") on node \"crc\" DevicePath \"\"" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.135042 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9b490b3-a462-4471-b9ea-ed125f0e4de8-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.135058 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9b490b3-a462-4471-b9ea-ed125f0e4de8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.482930 4808 generic.go:334] "Generic (PLEG): container finished" podID="b9b490b3-a462-4471-b9ea-ed125f0e4de8" containerID="84622f3612c0e8e1668729dadc088c0e107eef001151a69dd0daf8fad47031c7" exitCode=0 Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.482985 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c54tv" event={"ID":"b9b490b3-a462-4471-b9ea-ed125f0e4de8","Type":"ContainerDied","Data":"84622f3612c0e8e1668729dadc088c0e107eef001151a69dd0daf8fad47031c7"} Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.483069 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c54tv" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.483092 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c54tv" event={"ID":"b9b490b3-a462-4471-b9ea-ed125f0e4de8","Type":"ContainerDied","Data":"8ab900936615333ecc4c087a70a96c15db7cad1e69bf463746b4555ede1c25da"} Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.483117 4808 scope.go:117] "RemoveContainer" containerID="84622f3612c0e8e1668729dadc088c0e107eef001151a69dd0daf8fad47031c7" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.508130 4808 scope.go:117] "RemoveContainer" containerID="ce2a7a4cdaa1027d3932fd8b52c941ddcb6f864bd936a7c276c7f7686a8efd3b" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.541763 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c54tv"] Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.542134 4808 scope.go:117] "RemoveContainer" containerID="1c46a2cf91be07d96b0fe7fffe31f04cd18331ec7ea378a5a84d6ebc5f921ba5" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.557052 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c54tv"] Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.595051 4808 scope.go:117] "RemoveContainer" containerID="84622f3612c0e8e1668729dadc088c0e107eef001151a69dd0daf8fad47031c7" Nov 24 18:04:55 crc kubenswrapper[4808]: E1124 18:04:55.595547 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84622f3612c0e8e1668729dadc088c0e107eef001151a69dd0daf8fad47031c7\": container with ID starting with 84622f3612c0e8e1668729dadc088c0e107eef001151a69dd0daf8fad47031c7 not found: ID does not exist" containerID="84622f3612c0e8e1668729dadc088c0e107eef001151a69dd0daf8fad47031c7" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.595672 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84622f3612c0e8e1668729dadc088c0e107eef001151a69dd0daf8fad47031c7"} err="failed to get container status \"84622f3612c0e8e1668729dadc088c0e107eef001151a69dd0daf8fad47031c7\": rpc error: code = NotFound desc = could not find container \"84622f3612c0e8e1668729dadc088c0e107eef001151a69dd0daf8fad47031c7\": container with ID starting with 84622f3612c0e8e1668729dadc088c0e107eef001151a69dd0daf8fad47031c7 not found: ID does not exist" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.595778 4808 scope.go:117] "RemoveContainer" containerID="ce2a7a4cdaa1027d3932fd8b52c941ddcb6f864bd936a7c276c7f7686a8efd3b" Nov 24 18:04:55 crc kubenswrapper[4808]: E1124 18:04:55.596219 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce2a7a4cdaa1027d3932fd8b52c941ddcb6f864bd936a7c276c7f7686a8efd3b\": container with ID starting with ce2a7a4cdaa1027d3932fd8b52c941ddcb6f864bd936a7c276c7f7686a8efd3b not found: ID does not exist" containerID="ce2a7a4cdaa1027d3932fd8b52c941ddcb6f864bd936a7c276c7f7686a8efd3b" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.596250 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce2a7a4cdaa1027d3932fd8b52c941ddcb6f864bd936a7c276c7f7686a8efd3b"} err="failed to get container status \"ce2a7a4cdaa1027d3932fd8b52c941ddcb6f864bd936a7c276c7f7686a8efd3b\": rpc error: code = NotFound desc = could not find container \"ce2a7a4cdaa1027d3932fd8b52c941ddcb6f864bd936a7c276c7f7686a8efd3b\": container with ID starting with ce2a7a4cdaa1027d3932fd8b52c941ddcb6f864bd936a7c276c7f7686a8efd3b not found: ID does not exist" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.596269 4808 scope.go:117] "RemoveContainer" containerID="1c46a2cf91be07d96b0fe7fffe31f04cd18331ec7ea378a5a84d6ebc5f921ba5" Nov 24 18:04:55 crc kubenswrapper[4808]: E1124 18:04:55.597780 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c46a2cf91be07d96b0fe7fffe31f04cd18331ec7ea378a5a84d6ebc5f921ba5\": container with ID starting with 1c46a2cf91be07d96b0fe7fffe31f04cd18331ec7ea378a5a84d6ebc5f921ba5 not found: ID does not exist" containerID="1c46a2cf91be07d96b0fe7fffe31f04cd18331ec7ea378a5a84d6ebc5f921ba5" Nov 24 18:04:55 crc kubenswrapper[4808]: I1124 18:04:55.597802 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c46a2cf91be07d96b0fe7fffe31f04cd18331ec7ea378a5a84d6ebc5f921ba5"} err="failed to get container status \"1c46a2cf91be07d96b0fe7fffe31f04cd18331ec7ea378a5a84d6ebc5f921ba5\": rpc error: code = NotFound desc = could not find container \"1c46a2cf91be07d96b0fe7fffe31f04cd18331ec7ea378a5a84d6ebc5f921ba5\": container with ID starting with 1c46a2cf91be07d96b0fe7fffe31f04cd18331ec7ea378a5a84d6ebc5f921ba5 not found: ID does not exist" Nov 24 18:04:56 crc kubenswrapper[4808]: I1124 18:04:56.357806 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9b490b3-a462-4471-b9ea-ed125f0e4de8" path="/var/lib/kubelet/pods/b9b490b3-a462-4471-b9ea-ed125f0e4de8/volumes" Nov 24 18:05:06 crc kubenswrapper[4808]: I1124 18:05:06.523220 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:05:06 crc kubenswrapper[4808]: I1124 18:05:06.523877 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:05:33 crc kubenswrapper[4808]: I1124 18:05:33.832780 4808 generic.go:334] "Generic (PLEG): container finished" podID="e03cdbd6-8f86-468f-a798-961c6a4920d7" containerID="824e7f71e1ed99ae2cb4f76fd6903bf5c887d80e0f148be270f745a9711f8b7c" exitCode=0 Nov 24 18:05:33 crc kubenswrapper[4808]: I1124 18:05:33.832872 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" event={"ID":"e03cdbd6-8f86-468f-a798-961c6a4920d7","Type":"ContainerDied","Data":"824e7f71e1ed99ae2cb4f76fd6903bf5c887d80e0f148be270f745a9711f8b7c"} Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.286464 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.353915 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-extra-config-0\") pod \"e03cdbd6-8f86-468f-a798-961c6a4920d7\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.353965 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-inventory\") pod \"e03cdbd6-8f86-468f-a798-961c6a4920d7\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.354090 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-migration-ssh-key-0\") pod \"e03cdbd6-8f86-468f-a798-961c6a4920d7\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.354152 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-combined-ca-bundle\") pod \"e03cdbd6-8f86-468f-a798-961c6a4920d7\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.354240 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-cell1-compute-config-1\") pod \"e03cdbd6-8f86-468f-a798-961c6a4920d7\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.354404 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-cell1-compute-config-0\") pod \"e03cdbd6-8f86-468f-a798-961c6a4920d7\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.354441 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntw5l\" (UniqueName: \"kubernetes.io/projected/e03cdbd6-8f86-468f-a798-961c6a4920d7-kube-api-access-ntw5l\") pod \"e03cdbd6-8f86-468f-a798-961c6a4920d7\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.354487 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-migration-ssh-key-1\") pod \"e03cdbd6-8f86-468f-a798-961c6a4920d7\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.354513 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-ssh-key\") pod \"e03cdbd6-8f86-468f-a798-961c6a4920d7\" (UID: \"e03cdbd6-8f86-468f-a798-961c6a4920d7\") " Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.359663 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e03cdbd6-8f86-468f-a798-961c6a4920d7-kube-api-access-ntw5l" (OuterVolumeSpecName: "kube-api-access-ntw5l") pod "e03cdbd6-8f86-468f-a798-961c6a4920d7" (UID: "e03cdbd6-8f86-468f-a798-961c6a4920d7"). InnerVolumeSpecName "kube-api-access-ntw5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.368790 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "e03cdbd6-8f86-468f-a798-961c6a4920d7" (UID: "e03cdbd6-8f86-468f-a798-961c6a4920d7"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.383123 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "e03cdbd6-8f86-468f-a798-961c6a4920d7" (UID: "e03cdbd6-8f86-468f-a798-961c6a4920d7"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.383638 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e03cdbd6-8f86-468f-a798-961c6a4920d7" (UID: "e03cdbd6-8f86-468f-a798-961c6a4920d7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.383659 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "e03cdbd6-8f86-468f-a798-961c6a4920d7" (UID: "e03cdbd6-8f86-468f-a798-961c6a4920d7"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.387433 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "e03cdbd6-8f86-468f-a798-961c6a4920d7" (UID: "e03cdbd6-8f86-468f-a798-961c6a4920d7"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.387467 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "e03cdbd6-8f86-468f-a798-961c6a4920d7" (UID: "e03cdbd6-8f86-468f-a798-961c6a4920d7"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.397286 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "e03cdbd6-8f86-468f-a798-961c6a4920d7" (UID: "e03cdbd6-8f86-468f-a798-961c6a4920d7"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.410654 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-inventory" (OuterVolumeSpecName: "inventory") pod "e03cdbd6-8f86-468f-a798-961c6a4920d7" (UID: "e03cdbd6-8f86-468f-a798-961c6a4920d7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.457338 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntw5l\" (UniqueName: \"kubernetes.io/projected/e03cdbd6-8f86-468f-a798-961c6a4920d7-kube-api-access-ntw5l\") on node \"crc\" DevicePath \"\"" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.457641 4808 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.457652 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.457661 4808 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.457671 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.457703 4808 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.457714 4808 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.457723 4808 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.457734 4808 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e03cdbd6-8f86-468f-a798-961c6a4920d7-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.857681 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" event={"ID":"e03cdbd6-8f86-468f-a798-961c6a4920d7","Type":"ContainerDied","Data":"31fb9b7853097eddf93e8b7cf5f3b23b22147a34441b836dae8fee7100d72049"} Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.857732 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31fb9b7853097eddf93e8b7cf5f3b23b22147a34441b836dae8fee7100d72049" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.857795 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lfxmh" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.962799 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n"] Nov 24 18:05:35 crc kubenswrapper[4808]: E1124 18:05:35.963273 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e03cdbd6-8f86-468f-a798-961c6a4920d7" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.963292 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e03cdbd6-8f86-468f-a798-961c6a4920d7" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 18:05:35 crc kubenswrapper[4808]: E1124 18:05:35.963324 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b490b3-a462-4471-b9ea-ed125f0e4de8" containerName="registry-server" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.963332 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b490b3-a462-4471-b9ea-ed125f0e4de8" containerName="registry-server" Nov 24 18:05:35 crc kubenswrapper[4808]: E1124 18:05:35.963358 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b490b3-a462-4471-b9ea-ed125f0e4de8" containerName="extract-content" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.963365 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b490b3-a462-4471-b9ea-ed125f0e4de8" containerName="extract-content" Nov 24 18:05:35 crc kubenswrapper[4808]: E1124 18:05:35.963387 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b490b3-a462-4471-b9ea-ed125f0e4de8" containerName="extract-utilities" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.963394 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b490b3-a462-4471-b9ea-ed125f0e4de8" containerName="extract-utilities" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.963589 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e03cdbd6-8f86-468f-a798-961c6a4920d7" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.963614 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9b490b3-a462-4471-b9ea-ed125f0e4de8" containerName="registry-server" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.964368 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:35 crc kubenswrapper[4808]: W1124 18:05:35.969817 4808 reflector.go:561] object-"openstack"/"ceilometer-compute-config-data": failed to list *v1.Secret: secrets "ceilometer-compute-config-data" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.969873 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.969907 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 18:05:35 crc kubenswrapper[4808]: E1124 18:05:35.969880 4808 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"ceilometer-compute-config-data\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ceilometer-compute-config-data\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.971282 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rc5lv" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.971477 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 18:05:35 crc kubenswrapper[4808]: I1124 18:05:35.986288 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n"] Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.069466 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.069566 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.069628 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.069724 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.069821 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.069858 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.069945 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt4hk\" (UniqueName: \"kubernetes.io/projected/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-kube-api-access-tt4hk\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.171415 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.171549 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.171588 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.171664 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt4hk\" (UniqueName: \"kubernetes.io/projected/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-kube-api-access-tt4hk\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.171751 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.171782 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.171852 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.178924 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.180899 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.189519 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.202081 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt4hk\" (UniqueName: \"kubernetes.io/projected/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-kube-api-access-tt4hk\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.522877 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.522944 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.522993 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.523756 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f"} pod="openshift-machine-config-operator/machine-config-daemon-ht87w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.523824 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" containerID="cri-o://493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" gracePeriod=600 Nov 24 18:05:36 crc kubenswrapper[4808]: E1124 18:05:36.677165 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.867788 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" exitCode=0 Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.867828 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerDied","Data":"493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f"} Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.868233 4808 scope.go:117] "RemoveContainer" containerID="a01acbe1d0e5b11ae5eb995ee52c951450e7a150c9c344c5eacfa7fe2bda2db2" Nov 24 18:05:36 crc kubenswrapper[4808]: I1124 18:05:36.869150 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:05:36 crc kubenswrapper[4808]: E1124 18:05:36.871634 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:05:37 crc kubenswrapper[4808]: E1124 18:05:37.171963 4808 secret.go:188] Couldn't get secret openstack/ceilometer-compute-config-data: failed to sync secret cache: timed out waiting for the condition Nov 24 18:05:37 crc kubenswrapper[4808]: E1124 18:05:37.171983 4808 secret.go:188] Couldn't get secret openstack/ceilometer-compute-config-data: failed to sync secret cache: timed out waiting for the condition Nov 24 18:05:37 crc kubenswrapper[4808]: E1124 18:05:37.172049 4808 secret.go:188] Couldn't get secret openstack/ceilometer-compute-config-data: failed to sync secret cache: timed out waiting for the condition Nov 24 18:05:37 crc kubenswrapper[4808]: E1124 18:05:37.172059 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-1 podName:7f556606-72e2-4fc7-bb57-cc4bd96c4f9b nodeName:}" failed. No retries permitted until 2025-11-24 18:05:37.672032151 +0000 UTC m=+2330.269699953 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ceilometer-compute-config-data-1" (UniqueName: "kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-1") pod "telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" (UID: "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b") : failed to sync secret cache: timed out waiting for the condition Nov 24 18:05:37 crc kubenswrapper[4808]: E1124 18:05:37.172083 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-2 podName:7f556606-72e2-4fc7-bb57-cc4bd96c4f9b nodeName:}" failed. No retries permitted until 2025-11-24 18:05:37.672069932 +0000 UTC m=+2330.269737744 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ceilometer-compute-config-data-2" (UniqueName: "kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-2") pod "telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" (UID: "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b") : failed to sync secret cache: timed out waiting for the condition Nov 24 18:05:37 crc kubenswrapper[4808]: E1124 18:05:37.172098 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-0 podName:7f556606-72e2-4fc7-bb57-cc4bd96c4f9b nodeName:}" failed. No retries permitted until 2025-11-24 18:05:37.672091622 +0000 UTC m=+2330.269759414 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ceilometer-compute-config-data-0" (UniqueName: "kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-0") pod "telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" (UID: "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b") : failed to sync secret cache: timed out waiting for the condition Nov 24 18:05:37 crc kubenswrapper[4808]: I1124 18:05:37.195546 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 24 18:05:37 crc kubenswrapper[4808]: I1124 18:05:37.702477 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:37 crc kubenswrapper[4808]: I1124 18:05:37.702626 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:37 crc kubenswrapper[4808]: I1124 18:05:37.702660 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:37 crc kubenswrapper[4808]: I1124 18:05:37.707822 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:37 crc kubenswrapper[4808]: I1124 18:05:37.708063 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:37 crc kubenswrapper[4808]: I1124 18:05:37.708619 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kt77n\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:37 crc kubenswrapper[4808]: I1124 18:05:37.791051 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:05:38 crc kubenswrapper[4808]: I1124 18:05:38.114703 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n"] Nov 24 18:05:38 crc kubenswrapper[4808]: I1124 18:05:38.898129 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" event={"ID":"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b","Type":"ContainerStarted","Data":"a14dbff3c13aa551968f9c3bb8a215dac48e1e6834b40f757898cf26e414d8a0"} Nov 24 18:05:38 crc kubenswrapper[4808]: I1124 18:05:38.898935 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" event={"ID":"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b","Type":"ContainerStarted","Data":"eafc51f1023f3b508d744e48f3effcdd8586843dd8f4703467acdc556906a45c"} Nov 24 18:05:38 crc kubenswrapper[4808]: I1124 18:05:38.921652 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" podStartSLOduration=3.446873105 podStartE2EDuration="3.921634393s" podCreationTimestamp="2025-11-24 18:05:35 +0000 UTC" firstStartedPulling="2025-11-24 18:05:38.121414402 +0000 UTC m=+2330.719082204" lastFinishedPulling="2025-11-24 18:05:38.59617568 +0000 UTC m=+2331.193843492" observedRunningTime="2025-11-24 18:05:38.918616989 +0000 UTC m=+2331.516284791" watchObservedRunningTime="2025-11-24 18:05:38.921634393 +0000 UTC m=+2331.519302195" Nov 24 18:05:52 crc kubenswrapper[4808]: I1124 18:05:52.347532 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:05:52 crc kubenswrapper[4808]: E1124 18:05:52.348301 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:06:06 crc kubenswrapper[4808]: I1124 18:06:06.347388 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:06:06 crc kubenswrapper[4808]: E1124 18:06:06.348440 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:06:20 crc kubenswrapper[4808]: I1124 18:06:20.347378 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:06:20 crc kubenswrapper[4808]: E1124 18:06:20.348174 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:06:31 crc kubenswrapper[4808]: I1124 18:06:31.346979 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:06:31 crc kubenswrapper[4808]: E1124 18:06:31.347844 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:06:45 crc kubenswrapper[4808]: I1124 18:06:45.348060 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:06:45 crc kubenswrapper[4808]: E1124 18:06:45.349213 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:07:00 crc kubenswrapper[4808]: I1124 18:07:00.347935 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:07:00 crc kubenswrapper[4808]: E1124 18:07:00.348821 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:07:13 crc kubenswrapper[4808]: I1124 18:07:13.347396 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:07:13 crc kubenswrapper[4808]: E1124 18:07:13.348541 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:07:27 crc kubenswrapper[4808]: I1124 18:07:27.348202 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:07:27 crc kubenswrapper[4808]: E1124 18:07:27.349381 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:07:40 crc kubenswrapper[4808]: I1124 18:07:40.347848 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:07:40 crc kubenswrapper[4808]: E1124 18:07:40.348396 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:07:53 crc kubenswrapper[4808]: I1124 18:07:53.348332 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:07:53 crc kubenswrapper[4808]: E1124 18:07:53.349386 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:07:59 crc kubenswrapper[4808]: I1124 18:07:59.294275 4808 generic.go:334] "Generic (PLEG): container finished" podID="7f556606-72e2-4fc7-bb57-cc4bd96c4f9b" containerID="a14dbff3c13aa551968f9c3bb8a215dac48e1e6834b40f757898cf26e414d8a0" exitCode=0 Nov 24 18:07:59 crc kubenswrapper[4808]: I1124 18:07:59.294323 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" event={"ID":"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b","Type":"ContainerDied","Data":"a14dbff3c13aa551968f9c3bb8a215dac48e1e6834b40f757898cf26e414d8a0"} Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.709297 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.765493 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-telemetry-combined-ca-bundle\") pod \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.765913 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ssh-key\") pod \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.766045 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-1\") pod \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.766189 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt4hk\" (UniqueName: \"kubernetes.io/projected/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-kube-api-access-tt4hk\") pod \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.766327 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-0\") pod \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.766477 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-2\") pod \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.766592 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-inventory\") pod \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\" (UID: \"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b\") " Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.778355 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b" (UID: "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.788454 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-kube-api-access-tt4hk" (OuterVolumeSpecName: "kube-api-access-tt4hk") pod "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b" (UID: "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b"). InnerVolumeSpecName "kube-api-access-tt4hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.800002 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b" (UID: "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.800650 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b" (UID: "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.813880 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b" (UID: "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.816112 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-inventory" (OuterVolumeSpecName: "inventory") pod "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b" (UID: "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.820259 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b" (UID: "7f556606-72e2-4fc7-bb57-cc4bd96c4f9b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.869096 4808 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.869140 4808 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.869155 4808 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.869170 4808 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.869182 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.869194 4808 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 24 18:08:00 crc kubenswrapper[4808]: I1124 18:08:00.869209 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt4hk\" (UniqueName: \"kubernetes.io/projected/7f556606-72e2-4fc7-bb57-cc4bd96c4f9b-kube-api-access-tt4hk\") on node \"crc\" DevicePath \"\"" Nov 24 18:08:01 crc kubenswrapper[4808]: I1124 18:08:01.316500 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" event={"ID":"7f556606-72e2-4fc7-bb57-cc4bd96c4f9b","Type":"ContainerDied","Data":"eafc51f1023f3b508d744e48f3effcdd8586843dd8f4703467acdc556906a45c"} Nov 24 18:08:01 crc kubenswrapper[4808]: I1124 18:08:01.316860 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eafc51f1023f3b508d744e48f3effcdd8586843dd8f4703467acdc556906a45c" Nov 24 18:08:01 crc kubenswrapper[4808]: I1124 18:08:01.316712 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kt77n" Nov 24 18:08:04 crc kubenswrapper[4808]: I1124 18:08:04.347683 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:08:04 crc kubenswrapper[4808]: E1124 18:08:04.347966 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:08:19 crc kubenswrapper[4808]: I1124 18:08:19.347867 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:08:19 crc kubenswrapper[4808]: E1124 18:08:19.348705 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:08:30 crc kubenswrapper[4808]: I1124 18:08:30.349883 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:08:30 crc kubenswrapper[4808]: E1124 18:08:30.350829 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:08:43 crc kubenswrapper[4808]: I1124 18:08:43.348560 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:08:43 crc kubenswrapper[4808]: E1124 18:08:43.349960 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.051565 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 18:08:54 crc kubenswrapper[4808]: E1124 18:08:54.052797 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f556606-72e2-4fc7-bb57-cc4bd96c4f9b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.052822 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f556606-72e2-4fc7-bb57-cc4bd96c4f9b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.053161 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f556606-72e2-4fc7-bb57-cc4bd96c4f9b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.054194 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.059449 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.062125 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.071006 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.108539 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.108678 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-xhqjb" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.109344 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d014b1af-f795-43e4-af00-0a152d0cb945-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.109466 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d014b1af-f795-43e4-af00-0a152d0cb945-config-data\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.109513 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.211411 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d014b1af-f795-43e4-af00-0a152d0cb945-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.211880 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.211918 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.212007 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.212121 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g77gz\" (UniqueName: \"kubernetes.io/projected/d014b1af-f795-43e4-af00-0a152d0cb945-kube-api-access-g77gz\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.212228 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d014b1af-f795-43e4-af00-0a152d0cb945-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.212301 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d014b1af-f795-43e4-af00-0a152d0cb945-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.212378 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d014b1af-f795-43e4-af00-0a152d0cb945-config-data\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.212426 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.214308 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d014b1af-f795-43e4-af00-0a152d0cb945-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.214493 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d014b1af-f795-43e4-af00-0a152d0cb945-config-data\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.221882 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.317781 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d014b1af-f795-43e4-af00-0a152d0cb945-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.317882 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.317928 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.318124 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.318457 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g77gz\" (UniqueName: \"kubernetes.io/projected/d014b1af-f795-43e4-af00-0a152d0cb945-kube-api-access-g77gz\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.318663 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d014b1af-f795-43e4-af00-0a152d0cb945-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.320989 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d014b1af-f795-43e4-af00-0a152d0cb945-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.328447 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.329181 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d014b1af-f795-43e4-af00-0a152d0cb945-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.330151 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.342167 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.359135 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g77gz\" (UniqueName: \"kubernetes.io/projected/d014b1af-f795-43e4-af00-0a152d0cb945-kube-api-access-g77gz\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.374008 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.425919 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 18:08:54 crc kubenswrapper[4808]: W1124 18:08:54.905612 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd014b1af_f795_43e4_af00_0a152d0cb945.slice/crio-e9dca01a08450a29b1101326a77535ac2472c681db1fe0acd2eadfbe47bf3853 WatchSource:0}: Error finding container e9dca01a08450a29b1101326a77535ac2472c681db1fe0acd2eadfbe47bf3853: Status 404 returned error can't find the container with id e9dca01a08450a29b1101326a77535ac2472c681db1fe0acd2eadfbe47bf3853 Nov 24 18:08:54 crc kubenswrapper[4808]: I1124 18:08:54.917986 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 18:08:55 crc kubenswrapper[4808]: I1124 18:08:55.347009 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:08:55 crc kubenswrapper[4808]: E1124 18:08:55.347599 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:08:55 crc kubenswrapper[4808]: I1124 18:08:55.887424 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d014b1af-f795-43e4-af00-0a152d0cb945","Type":"ContainerStarted","Data":"e9dca01a08450a29b1101326a77535ac2472c681db1fe0acd2eadfbe47bf3853"} Nov 24 18:09:10 crc kubenswrapper[4808]: I1124 18:09:10.348815 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:09:10 crc kubenswrapper[4808]: E1124 18:09:10.349476 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:09:21 crc kubenswrapper[4808]: E1124 18:09:21.982859 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 24 18:09:21 crc kubenswrapper[4808]: E1124 18:09:21.984377 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g77gz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(d014b1af-f795-43e4-af00-0a152d0cb945): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 18:09:21 crc kubenswrapper[4808]: E1124 18:09:21.985678 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="d014b1af-f795-43e4-af00-0a152d0cb945" Nov 24 18:09:22 crc kubenswrapper[4808]: E1124 18:09:22.133182 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="d014b1af-f795-43e4-af00-0a152d0cb945" Nov 24 18:09:23 crc kubenswrapper[4808]: I1124 18:09:23.347841 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:09:23 crc kubenswrapper[4808]: E1124 18:09:23.348547 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:09:37 crc kubenswrapper[4808]: I1124 18:09:37.276982 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d014b1af-f795-43e4-af00-0a152d0cb945","Type":"ContainerStarted","Data":"72d4294f444f4381cf5070cb811ee1e5eb4f4c6b3578bdf87b2a75cd3a203826"} Nov 24 18:09:37 crc kubenswrapper[4808]: I1124 18:09:37.301815 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.374211357 podStartE2EDuration="44.301798784s" podCreationTimestamp="2025-11-24 18:08:53 +0000 UTC" firstStartedPulling="2025-11-24 18:08:54.908458164 +0000 UTC m=+2527.506125976" lastFinishedPulling="2025-11-24 18:09:35.836045561 +0000 UTC m=+2568.433713403" observedRunningTime="2025-11-24 18:09:37.297205526 +0000 UTC m=+2569.894873338" watchObservedRunningTime="2025-11-24 18:09:37.301798784 +0000 UTC m=+2569.899466586" Nov 24 18:09:37 crc kubenswrapper[4808]: I1124 18:09:37.347815 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:09:37 crc kubenswrapper[4808]: E1124 18:09:37.348143 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:09:52 crc kubenswrapper[4808]: I1124 18:09:52.347315 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:09:52 crc kubenswrapper[4808]: E1124 18:09:52.348244 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:10:04 crc kubenswrapper[4808]: I1124 18:10:04.347586 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:10:04 crc kubenswrapper[4808]: E1124 18:10:04.348341 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:10:18 crc kubenswrapper[4808]: I1124 18:10:18.353136 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:10:18 crc kubenswrapper[4808]: E1124 18:10:18.354139 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:10:29 crc kubenswrapper[4808]: I1124 18:10:29.347421 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:10:29 crc kubenswrapper[4808]: E1124 18:10:29.348083 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:10:43 crc kubenswrapper[4808]: I1124 18:10:43.348004 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:10:43 crc kubenswrapper[4808]: I1124 18:10:43.996090 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"40b4e634cbb4a00624c13f81bfdf705c343a7ea5c4f0bdc8a8675135a0436bfa"} Nov 24 18:11:15 crc kubenswrapper[4808]: I1124 18:11:15.262667 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8kmz9"] Nov 24 18:11:15 crc kubenswrapper[4808]: I1124 18:11:15.305387 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8kmz9" Nov 24 18:11:15 crc kubenswrapper[4808]: I1124 18:11:15.311169 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8kmz9"] Nov 24 18:11:15 crc kubenswrapper[4808]: I1124 18:11:15.509048 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13f73b0-c183-4f29-b75d-2fef31356a61-catalog-content\") pod \"redhat-operators-8kmz9\" (UID: \"c13f73b0-c183-4f29-b75d-2fef31356a61\") " pod="openshift-marketplace/redhat-operators-8kmz9" Nov 24 18:11:15 crc kubenswrapper[4808]: I1124 18:11:15.509300 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jphlg\" (UniqueName: \"kubernetes.io/projected/c13f73b0-c183-4f29-b75d-2fef31356a61-kube-api-access-jphlg\") pod \"redhat-operators-8kmz9\" (UID: \"c13f73b0-c183-4f29-b75d-2fef31356a61\") " pod="openshift-marketplace/redhat-operators-8kmz9" Nov 24 18:11:15 crc kubenswrapper[4808]: I1124 18:11:15.509623 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13f73b0-c183-4f29-b75d-2fef31356a61-utilities\") pod \"redhat-operators-8kmz9\" (UID: \"c13f73b0-c183-4f29-b75d-2fef31356a61\") " pod="openshift-marketplace/redhat-operators-8kmz9" Nov 24 18:11:15 crc kubenswrapper[4808]: I1124 18:11:15.611765 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13f73b0-c183-4f29-b75d-2fef31356a61-utilities\") pod \"redhat-operators-8kmz9\" (UID: \"c13f73b0-c183-4f29-b75d-2fef31356a61\") " pod="openshift-marketplace/redhat-operators-8kmz9" Nov 24 18:11:15 crc kubenswrapper[4808]: I1124 18:11:15.611915 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13f73b0-c183-4f29-b75d-2fef31356a61-catalog-content\") pod \"redhat-operators-8kmz9\" (UID: \"c13f73b0-c183-4f29-b75d-2fef31356a61\") " pod="openshift-marketplace/redhat-operators-8kmz9" Nov 24 18:11:15 crc kubenswrapper[4808]: I1124 18:11:15.612008 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jphlg\" (UniqueName: \"kubernetes.io/projected/c13f73b0-c183-4f29-b75d-2fef31356a61-kube-api-access-jphlg\") pod \"redhat-operators-8kmz9\" (UID: \"c13f73b0-c183-4f29-b75d-2fef31356a61\") " pod="openshift-marketplace/redhat-operators-8kmz9" Nov 24 18:11:15 crc kubenswrapper[4808]: I1124 18:11:15.612523 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13f73b0-c183-4f29-b75d-2fef31356a61-utilities\") pod \"redhat-operators-8kmz9\" (UID: \"c13f73b0-c183-4f29-b75d-2fef31356a61\") " pod="openshift-marketplace/redhat-operators-8kmz9" Nov 24 18:11:15 crc kubenswrapper[4808]: I1124 18:11:15.612578 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13f73b0-c183-4f29-b75d-2fef31356a61-catalog-content\") pod \"redhat-operators-8kmz9\" (UID: \"c13f73b0-c183-4f29-b75d-2fef31356a61\") " pod="openshift-marketplace/redhat-operators-8kmz9" Nov 24 18:11:15 crc kubenswrapper[4808]: I1124 18:11:15.637454 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jphlg\" (UniqueName: \"kubernetes.io/projected/c13f73b0-c183-4f29-b75d-2fef31356a61-kube-api-access-jphlg\") pod \"redhat-operators-8kmz9\" (UID: \"c13f73b0-c183-4f29-b75d-2fef31356a61\") " pod="openshift-marketplace/redhat-operators-8kmz9" Nov 24 18:11:15 crc kubenswrapper[4808]: I1124 18:11:15.936847 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8kmz9" Nov 24 18:11:16 crc kubenswrapper[4808]: I1124 18:11:16.440443 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8kmz9"] Nov 24 18:11:17 crc kubenswrapper[4808]: I1124 18:11:17.370706 4808 generic.go:334] "Generic (PLEG): container finished" podID="c13f73b0-c183-4f29-b75d-2fef31356a61" containerID="42758ab8c0335118e9a6a0bcd605fe92b65e4500e5f637a85d181cda070100e9" exitCode=0 Nov 24 18:11:17 crc kubenswrapper[4808]: I1124 18:11:17.370934 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8kmz9" event={"ID":"c13f73b0-c183-4f29-b75d-2fef31356a61","Type":"ContainerDied","Data":"42758ab8c0335118e9a6a0bcd605fe92b65e4500e5f637a85d181cda070100e9"} Nov 24 18:11:17 crc kubenswrapper[4808]: I1124 18:11:17.371230 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8kmz9" event={"ID":"c13f73b0-c183-4f29-b75d-2fef31356a61","Type":"ContainerStarted","Data":"7707b864fc148e42ea62fae5ed2b8e6e906f6496615c7a5997a6ec3dccb7c5c5"} Nov 24 18:11:17 crc kubenswrapper[4808]: I1124 18:11:17.373603 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 18:11:26 crc kubenswrapper[4808]: I1124 18:11:26.469148 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8kmz9" event={"ID":"c13f73b0-c183-4f29-b75d-2fef31356a61","Type":"ContainerStarted","Data":"34f936b29601e2bbfee81e8d19e0447bbdf11918839acdafcb7224e070b24ef7"} Nov 24 18:11:27 crc kubenswrapper[4808]: I1124 18:11:27.483364 4808 generic.go:334] "Generic (PLEG): container finished" podID="c13f73b0-c183-4f29-b75d-2fef31356a61" containerID="34f936b29601e2bbfee81e8d19e0447bbdf11918839acdafcb7224e070b24ef7" exitCode=0 Nov 24 18:11:27 crc kubenswrapper[4808]: I1124 18:11:27.483426 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8kmz9" event={"ID":"c13f73b0-c183-4f29-b75d-2fef31356a61","Type":"ContainerDied","Data":"34f936b29601e2bbfee81e8d19e0447bbdf11918839acdafcb7224e070b24ef7"} Nov 24 18:11:29 crc kubenswrapper[4808]: I1124 18:11:29.505654 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8kmz9" event={"ID":"c13f73b0-c183-4f29-b75d-2fef31356a61","Type":"ContainerStarted","Data":"ab7929505b9e357fc2ef6315aefb6d923676da8b09199cae0b7d3fc6436780bd"} Nov 24 18:11:29 crc kubenswrapper[4808]: I1124 18:11:29.527933 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8kmz9" podStartSLOduration=2.986347012 podStartE2EDuration="14.527877039s" podCreationTimestamp="2025-11-24 18:11:15 +0000 UTC" firstStartedPulling="2025-11-24 18:11:17.373225706 +0000 UTC m=+2669.970893548" lastFinishedPulling="2025-11-24 18:11:28.914755753 +0000 UTC m=+2681.512423575" observedRunningTime="2025-11-24 18:11:29.525141573 +0000 UTC m=+2682.122809375" watchObservedRunningTime="2025-11-24 18:11:29.527877039 +0000 UTC m=+2682.125544871" Nov 24 18:11:35 crc kubenswrapper[4808]: I1124 18:11:35.937002 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8kmz9" Nov 24 18:11:35 crc kubenswrapper[4808]: I1124 18:11:35.937494 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8kmz9" Nov 24 18:11:35 crc kubenswrapper[4808]: I1124 18:11:35.999727 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8kmz9" Nov 24 18:11:36 crc kubenswrapper[4808]: I1124 18:11:36.621852 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8kmz9" Nov 24 18:11:36 crc kubenswrapper[4808]: I1124 18:11:36.720333 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8kmz9"] Nov 24 18:11:36 crc kubenswrapper[4808]: I1124 18:11:36.758519 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cnl6r"] Nov 24 18:11:36 crc kubenswrapper[4808]: I1124 18:11:36.758802 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cnl6r" podUID="c935237c-9ccc-49e7-81b8-c770dcc3665b" containerName="registry-server" containerID="cri-o://f596ec07b1f76054fd189d12901fca0476be5d8f6f131e5740c03581cbeeba8d" gracePeriod=2 Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.317997 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.358443 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjlk9\" (UniqueName: \"kubernetes.io/projected/c935237c-9ccc-49e7-81b8-c770dcc3665b-kube-api-access-pjlk9\") pod \"c935237c-9ccc-49e7-81b8-c770dcc3665b\" (UID: \"c935237c-9ccc-49e7-81b8-c770dcc3665b\") " Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.358520 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c935237c-9ccc-49e7-81b8-c770dcc3665b-utilities\") pod \"c935237c-9ccc-49e7-81b8-c770dcc3665b\" (UID: \"c935237c-9ccc-49e7-81b8-c770dcc3665b\") " Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.365244 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c935237c-9ccc-49e7-81b8-c770dcc3665b-utilities" (OuterVolumeSpecName: "utilities") pod "c935237c-9ccc-49e7-81b8-c770dcc3665b" (UID: "c935237c-9ccc-49e7-81b8-c770dcc3665b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.370669 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c935237c-9ccc-49e7-81b8-c770dcc3665b-kube-api-access-pjlk9" (OuterVolumeSpecName: "kube-api-access-pjlk9") pod "c935237c-9ccc-49e7-81b8-c770dcc3665b" (UID: "c935237c-9ccc-49e7-81b8-c770dcc3665b"). InnerVolumeSpecName "kube-api-access-pjlk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.459953 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c935237c-9ccc-49e7-81b8-c770dcc3665b-catalog-content\") pod \"c935237c-9ccc-49e7-81b8-c770dcc3665b\" (UID: \"c935237c-9ccc-49e7-81b8-c770dcc3665b\") " Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.460691 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjlk9\" (UniqueName: \"kubernetes.io/projected/c935237c-9ccc-49e7-81b8-c770dcc3665b-kube-api-access-pjlk9\") on node \"crc\" DevicePath \"\"" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.460720 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c935237c-9ccc-49e7-81b8-c770dcc3665b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.576089 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c935237c-9ccc-49e7-81b8-c770dcc3665b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c935237c-9ccc-49e7-81b8-c770dcc3665b" (UID: "c935237c-9ccc-49e7-81b8-c770dcc3665b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.585420 4808 generic.go:334] "Generic (PLEG): container finished" podID="c935237c-9ccc-49e7-81b8-c770dcc3665b" containerID="f596ec07b1f76054fd189d12901fca0476be5d8f6f131e5740c03581cbeeba8d" exitCode=0 Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.586536 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cnl6r" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.589062 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cnl6r" event={"ID":"c935237c-9ccc-49e7-81b8-c770dcc3665b","Type":"ContainerDied","Data":"f596ec07b1f76054fd189d12901fca0476be5d8f6f131e5740c03581cbeeba8d"} Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.589098 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cnl6r" event={"ID":"c935237c-9ccc-49e7-81b8-c770dcc3665b","Type":"ContainerDied","Data":"6d5cd38e4f0ad115fe9191b1ea908f7821c876dd5957e88890e0f31813c3ecd1"} Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.589116 4808 scope.go:117] "RemoveContainer" containerID="f596ec07b1f76054fd189d12901fca0476be5d8f6f131e5740c03581cbeeba8d" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.617002 4808 scope.go:117] "RemoveContainer" containerID="582d3ed297d4554f523e547b529107bd21accfd72fe0c71ce521531b0c7210ca" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.626129 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cnl6r"] Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.635806 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cnl6r"] Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.646133 4808 scope.go:117] "RemoveContainer" containerID="eb216262205a74dafc233cd764d420cbfac697565bfdded8259afc744dad1608" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.664476 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c935237c-9ccc-49e7-81b8-c770dcc3665b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.684160 4808 scope.go:117] "RemoveContainer" containerID="f596ec07b1f76054fd189d12901fca0476be5d8f6f131e5740c03581cbeeba8d" Nov 24 18:11:37 crc kubenswrapper[4808]: E1124 18:11:37.684508 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f596ec07b1f76054fd189d12901fca0476be5d8f6f131e5740c03581cbeeba8d\": container with ID starting with f596ec07b1f76054fd189d12901fca0476be5d8f6f131e5740c03581cbeeba8d not found: ID does not exist" containerID="f596ec07b1f76054fd189d12901fca0476be5d8f6f131e5740c03581cbeeba8d" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.684534 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f596ec07b1f76054fd189d12901fca0476be5d8f6f131e5740c03581cbeeba8d"} err="failed to get container status \"f596ec07b1f76054fd189d12901fca0476be5d8f6f131e5740c03581cbeeba8d\": rpc error: code = NotFound desc = could not find container \"f596ec07b1f76054fd189d12901fca0476be5d8f6f131e5740c03581cbeeba8d\": container with ID starting with f596ec07b1f76054fd189d12901fca0476be5d8f6f131e5740c03581cbeeba8d not found: ID does not exist" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.684556 4808 scope.go:117] "RemoveContainer" containerID="582d3ed297d4554f523e547b529107bd21accfd72fe0c71ce521531b0c7210ca" Nov 24 18:11:37 crc kubenswrapper[4808]: E1124 18:11:37.684772 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"582d3ed297d4554f523e547b529107bd21accfd72fe0c71ce521531b0c7210ca\": container with ID starting with 582d3ed297d4554f523e547b529107bd21accfd72fe0c71ce521531b0c7210ca not found: ID does not exist" containerID="582d3ed297d4554f523e547b529107bd21accfd72fe0c71ce521531b0c7210ca" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.684795 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"582d3ed297d4554f523e547b529107bd21accfd72fe0c71ce521531b0c7210ca"} err="failed to get container status \"582d3ed297d4554f523e547b529107bd21accfd72fe0c71ce521531b0c7210ca\": rpc error: code = NotFound desc = could not find container \"582d3ed297d4554f523e547b529107bd21accfd72fe0c71ce521531b0c7210ca\": container with ID starting with 582d3ed297d4554f523e547b529107bd21accfd72fe0c71ce521531b0c7210ca not found: ID does not exist" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.684816 4808 scope.go:117] "RemoveContainer" containerID="eb216262205a74dafc233cd764d420cbfac697565bfdded8259afc744dad1608" Nov 24 18:11:37 crc kubenswrapper[4808]: E1124 18:11:37.685083 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb216262205a74dafc233cd764d420cbfac697565bfdded8259afc744dad1608\": container with ID starting with eb216262205a74dafc233cd764d420cbfac697565bfdded8259afc744dad1608 not found: ID does not exist" containerID="eb216262205a74dafc233cd764d420cbfac697565bfdded8259afc744dad1608" Nov 24 18:11:37 crc kubenswrapper[4808]: I1124 18:11:37.685104 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb216262205a74dafc233cd764d420cbfac697565bfdded8259afc744dad1608"} err="failed to get container status \"eb216262205a74dafc233cd764d420cbfac697565bfdded8259afc744dad1608\": rpc error: code = NotFound desc = could not find container \"eb216262205a74dafc233cd764d420cbfac697565bfdded8259afc744dad1608\": container with ID starting with eb216262205a74dafc233cd764d420cbfac697565bfdded8259afc744dad1608 not found: ID does not exist" Nov 24 18:11:38 crc kubenswrapper[4808]: I1124 18:11:38.358069 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c935237c-9ccc-49e7-81b8-c770dcc3665b" path="/var/lib/kubelet/pods/c935237c-9ccc-49e7-81b8-c770dcc3665b/volumes" Nov 24 18:11:41 crc kubenswrapper[4808]: I1124 18:11:41.925521 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kxwc5"] Nov 24 18:11:41 crc kubenswrapper[4808]: E1124 18:11:41.929553 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c935237c-9ccc-49e7-81b8-c770dcc3665b" containerName="extract-content" Nov 24 18:11:41 crc kubenswrapper[4808]: I1124 18:11:41.929576 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c935237c-9ccc-49e7-81b8-c770dcc3665b" containerName="extract-content" Nov 24 18:11:41 crc kubenswrapper[4808]: E1124 18:11:41.929592 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c935237c-9ccc-49e7-81b8-c770dcc3665b" containerName="registry-server" Nov 24 18:11:41 crc kubenswrapper[4808]: I1124 18:11:41.929599 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c935237c-9ccc-49e7-81b8-c770dcc3665b" containerName="registry-server" Nov 24 18:11:41 crc kubenswrapper[4808]: E1124 18:11:41.929623 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c935237c-9ccc-49e7-81b8-c770dcc3665b" containerName="extract-utilities" Nov 24 18:11:41 crc kubenswrapper[4808]: I1124 18:11:41.929630 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c935237c-9ccc-49e7-81b8-c770dcc3665b" containerName="extract-utilities" Nov 24 18:11:41 crc kubenswrapper[4808]: I1124 18:11:41.929877 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c935237c-9ccc-49e7-81b8-c770dcc3665b" containerName="registry-server" Nov 24 18:11:41 crc kubenswrapper[4808]: I1124 18:11:41.931539 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:41 crc kubenswrapper[4808]: I1124 18:11:41.949203 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kxwc5"] Nov 24 18:11:42 crc kubenswrapper[4808]: I1124 18:11:42.056534 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-475gt\" (UniqueName: \"kubernetes.io/projected/99e39212-efe7-46b3-a03b-77d7110ade81-kube-api-access-475gt\") pod \"community-operators-kxwc5\" (UID: \"99e39212-efe7-46b3-a03b-77d7110ade81\") " pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:42 crc kubenswrapper[4808]: I1124 18:11:42.056703 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99e39212-efe7-46b3-a03b-77d7110ade81-catalog-content\") pod \"community-operators-kxwc5\" (UID: \"99e39212-efe7-46b3-a03b-77d7110ade81\") " pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:42 crc kubenswrapper[4808]: I1124 18:11:42.056772 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99e39212-efe7-46b3-a03b-77d7110ade81-utilities\") pod \"community-operators-kxwc5\" (UID: \"99e39212-efe7-46b3-a03b-77d7110ade81\") " pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:42 crc kubenswrapper[4808]: I1124 18:11:42.158323 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99e39212-efe7-46b3-a03b-77d7110ade81-catalog-content\") pod \"community-operators-kxwc5\" (UID: \"99e39212-efe7-46b3-a03b-77d7110ade81\") " pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:42 crc kubenswrapper[4808]: I1124 18:11:42.158401 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99e39212-efe7-46b3-a03b-77d7110ade81-utilities\") pod \"community-operators-kxwc5\" (UID: \"99e39212-efe7-46b3-a03b-77d7110ade81\") " pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:42 crc kubenswrapper[4808]: I1124 18:11:42.158477 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-475gt\" (UniqueName: \"kubernetes.io/projected/99e39212-efe7-46b3-a03b-77d7110ade81-kube-api-access-475gt\") pod \"community-operators-kxwc5\" (UID: \"99e39212-efe7-46b3-a03b-77d7110ade81\") " pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:42 crc kubenswrapper[4808]: I1124 18:11:42.159525 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99e39212-efe7-46b3-a03b-77d7110ade81-catalog-content\") pod \"community-operators-kxwc5\" (UID: \"99e39212-efe7-46b3-a03b-77d7110ade81\") " pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:42 crc kubenswrapper[4808]: I1124 18:11:42.159808 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99e39212-efe7-46b3-a03b-77d7110ade81-utilities\") pod \"community-operators-kxwc5\" (UID: \"99e39212-efe7-46b3-a03b-77d7110ade81\") " pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:42 crc kubenswrapper[4808]: I1124 18:11:42.182870 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-475gt\" (UniqueName: \"kubernetes.io/projected/99e39212-efe7-46b3-a03b-77d7110ade81-kube-api-access-475gt\") pod \"community-operators-kxwc5\" (UID: \"99e39212-efe7-46b3-a03b-77d7110ade81\") " pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:42 crc kubenswrapper[4808]: I1124 18:11:42.260353 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:42 crc kubenswrapper[4808]: I1124 18:11:42.831057 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kxwc5"] Nov 24 18:11:43 crc kubenswrapper[4808]: I1124 18:11:43.644984 4808 generic.go:334] "Generic (PLEG): container finished" podID="99e39212-efe7-46b3-a03b-77d7110ade81" containerID="8fda896714711d9ce9687adc74a30b992983e810fbe4fe7f455e82b2428be8f2" exitCode=0 Nov 24 18:11:43 crc kubenswrapper[4808]: I1124 18:11:43.645081 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxwc5" event={"ID":"99e39212-efe7-46b3-a03b-77d7110ade81","Type":"ContainerDied","Data":"8fda896714711d9ce9687adc74a30b992983e810fbe4fe7f455e82b2428be8f2"} Nov 24 18:11:43 crc kubenswrapper[4808]: I1124 18:11:43.647439 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxwc5" event={"ID":"99e39212-efe7-46b3-a03b-77d7110ade81","Type":"ContainerStarted","Data":"27add8fa1665896d90883d9d1feeb10bedf2c3943b1c6dc2a4b9bc1604ca1910"} Nov 24 18:11:44 crc kubenswrapper[4808]: I1124 18:11:44.658826 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxwc5" event={"ID":"99e39212-efe7-46b3-a03b-77d7110ade81","Type":"ContainerStarted","Data":"c94e75dd8fda63437be4128c8388626a86fcf81760fb9a770e2be97e7cbdb15e"} Nov 24 18:11:45 crc kubenswrapper[4808]: I1124 18:11:45.670617 4808 generic.go:334] "Generic (PLEG): container finished" podID="99e39212-efe7-46b3-a03b-77d7110ade81" containerID="c94e75dd8fda63437be4128c8388626a86fcf81760fb9a770e2be97e7cbdb15e" exitCode=0 Nov 24 18:11:45 crc kubenswrapper[4808]: I1124 18:11:45.670698 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxwc5" event={"ID":"99e39212-efe7-46b3-a03b-77d7110ade81","Type":"ContainerDied","Data":"c94e75dd8fda63437be4128c8388626a86fcf81760fb9a770e2be97e7cbdb15e"} Nov 24 18:11:47 crc kubenswrapper[4808]: I1124 18:11:47.695647 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxwc5" event={"ID":"99e39212-efe7-46b3-a03b-77d7110ade81","Type":"ContainerStarted","Data":"64b190f46dd16016d980d44727177d7768f0aeaa20d5d239743c07c777772d20"} Nov 24 18:11:47 crc kubenswrapper[4808]: I1124 18:11:47.720823 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kxwc5" podStartSLOduration=3.7537313340000003 podStartE2EDuration="6.720801853s" podCreationTimestamp="2025-11-24 18:11:41 +0000 UTC" firstStartedPulling="2025-11-24 18:11:43.6475656 +0000 UTC m=+2696.245233402" lastFinishedPulling="2025-11-24 18:11:46.614636089 +0000 UTC m=+2699.212303921" observedRunningTime="2025-11-24 18:11:47.713418926 +0000 UTC m=+2700.311086728" watchObservedRunningTime="2025-11-24 18:11:47.720801853 +0000 UTC m=+2700.318469655" Nov 24 18:11:52 crc kubenswrapper[4808]: I1124 18:11:52.260834 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:52 crc kubenswrapper[4808]: I1124 18:11:52.263220 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:52 crc kubenswrapper[4808]: I1124 18:11:52.338913 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:52 crc kubenswrapper[4808]: I1124 18:11:52.817973 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:52 crc kubenswrapper[4808]: I1124 18:11:52.870808 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kxwc5"] Nov 24 18:11:54 crc kubenswrapper[4808]: I1124 18:11:54.759969 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kxwc5" podUID="99e39212-efe7-46b3-a03b-77d7110ade81" containerName="registry-server" containerID="cri-o://64b190f46dd16016d980d44727177d7768f0aeaa20d5d239743c07c777772d20" gracePeriod=2 Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.229348 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.319623 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99e39212-efe7-46b3-a03b-77d7110ade81-catalog-content\") pod \"99e39212-efe7-46b3-a03b-77d7110ade81\" (UID: \"99e39212-efe7-46b3-a03b-77d7110ade81\") " Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.319678 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-475gt\" (UniqueName: \"kubernetes.io/projected/99e39212-efe7-46b3-a03b-77d7110ade81-kube-api-access-475gt\") pod \"99e39212-efe7-46b3-a03b-77d7110ade81\" (UID: \"99e39212-efe7-46b3-a03b-77d7110ade81\") " Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.320522 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99e39212-efe7-46b3-a03b-77d7110ade81-utilities\") pod \"99e39212-efe7-46b3-a03b-77d7110ade81\" (UID: \"99e39212-efe7-46b3-a03b-77d7110ade81\") " Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.321451 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99e39212-efe7-46b3-a03b-77d7110ade81-utilities" (OuterVolumeSpecName: "utilities") pod "99e39212-efe7-46b3-a03b-77d7110ade81" (UID: "99e39212-efe7-46b3-a03b-77d7110ade81"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.329078 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99e39212-efe7-46b3-a03b-77d7110ade81-kube-api-access-475gt" (OuterVolumeSpecName: "kube-api-access-475gt") pod "99e39212-efe7-46b3-a03b-77d7110ade81" (UID: "99e39212-efe7-46b3-a03b-77d7110ade81"). InnerVolumeSpecName "kube-api-access-475gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.375722 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99e39212-efe7-46b3-a03b-77d7110ade81-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "99e39212-efe7-46b3-a03b-77d7110ade81" (UID: "99e39212-efe7-46b3-a03b-77d7110ade81"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.423223 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99e39212-efe7-46b3-a03b-77d7110ade81-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.423253 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-475gt\" (UniqueName: \"kubernetes.io/projected/99e39212-efe7-46b3-a03b-77d7110ade81-kube-api-access-475gt\") on node \"crc\" DevicePath \"\"" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.423264 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99e39212-efe7-46b3-a03b-77d7110ade81-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.772313 4808 generic.go:334] "Generic (PLEG): container finished" podID="99e39212-efe7-46b3-a03b-77d7110ade81" containerID="64b190f46dd16016d980d44727177d7768f0aeaa20d5d239743c07c777772d20" exitCode=0 Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.772500 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxwc5" event={"ID":"99e39212-efe7-46b3-a03b-77d7110ade81","Type":"ContainerDied","Data":"64b190f46dd16016d980d44727177d7768f0aeaa20d5d239743c07c777772d20"} Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.772727 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxwc5" event={"ID":"99e39212-efe7-46b3-a03b-77d7110ade81","Type":"ContainerDied","Data":"27add8fa1665896d90883d9d1feeb10bedf2c3943b1c6dc2a4b9bc1604ca1910"} Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.772762 4808 scope.go:117] "RemoveContainer" containerID="64b190f46dd16016d980d44727177d7768f0aeaa20d5d239743c07c777772d20" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.772598 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxwc5" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.807062 4808 scope.go:117] "RemoveContainer" containerID="c94e75dd8fda63437be4128c8388626a86fcf81760fb9a770e2be97e7cbdb15e" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.820463 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kxwc5"] Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.833140 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kxwc5"] Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.842756 4808 scope.go:117] "RemoveContainer" containerID="8fda896714711d9ce9687adc74a30b992983e810fbe4fe7f455e82b2428be8f2" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.879609 4808 scope.go:117] "RemoveContainer" containerID="64b190f46dd16016d980d44727177d7768f0aeaa20d5d239743c07c777772d20" Nov 24 18:11:55 crc kubenswrapper[4808]: E1124 18:11:55.880096 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64b190f46dd16016d980d44727177d7768f0aeaa20d5d239743c07c777772d20\": container with ID starting with 64b190f46dd16016d980d44727177d7768f0aeaa20d5d239743c07c777772d20 not found: ID does not exist" containerID="64b190f46dd16016d980d44727177d7768f0aeaa20d5d239743c07c777772d20" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.880171 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64b190f46dd16016d980d44727177d7768f0aeaa20d5d239743c07c777772d20"} err="failed to get container status \"64b190f46dd16016d980d44727177d7768f0aeaa20d5d239743c07c777772d20\": rpc error: code = NotFound desc = could not find container \"64b190f46dd16016d980d44727177d7768f0aeaa20d5d239743c07c777772d20\": container with ID starting with 64b190f46dd16016d980d44727177d7768f0aeaa20d5d239743c07c777772d20 not found: ID does not exist" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.880223 4808 scope.go:117] "RemoveContainer" containerID="c94e75dd8fda63437be4128c8388626a86fcf81760fb9a770e2be97e7cbdb15e" Nov 24 18:11:55 crc kubenswrapper[4808]: E1124 18:11:55.880641 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c94e75dd8fda63437be4128c8388626a86fcf81760fb9a770e2be97e7cbdb15e\": container with ID starting with c94e75dd8fda63437be4128c8388626a86fcf81760fb9a770e2be97e7cbdb15e not found: ID does not exist" containerID="c94e75dd8fda63437be4128c8388626a86fcf81760fb9a770e2be97e7cbdb15e" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.880683 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c94e75dd8fda63437be4128c8388626a86fcf81760fb9a770e2be97e7cbdb15e"} err="failed to get container status \"c94e75dd8fda63437be4128c8388626a86fcf81760fb9a770e2be97e7cbdb15e\": rpc error: code = NotFound desc = could not find container \"c94e75dd8fda63437be4128c8388626a86fcf81760fb9a770e2be97e7cbdb15e\": container with ID starting with c94e75dd8fda63437be4128c8388626a86fcf81760fb9a770e2be97e7cbdb15e not found: ID does not exist" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.880710 4808 scope.go:117] "RemoveContainer" containerID="8fda896714711d9ce9687adc74a30b992983e810fbe4fe7f455e82b2428be8f2" Nov 24 18:11:55 crc kubenswrapper[4808]: E1124 18:11:55.881000 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fda896714711d9ce9687adc74a30b992983e810fbe4fe7f455e82b2428be8f2\": container with ID starting with 8fda896714711d9ce9687adc74a30b992983e810fbe4fe7f455e82b2428be8f2 not found: ID does not exist" containerID="8fda896714711d9ce9687adc74a30b992983e810fbe4fe7f455e82b2428be8f2" Nov 24 18:11:55 crc kubenswrapper[4808]: I1124 18:11:55.881047 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fda896714711d9ce9687adc74a30b992983e810fbe4fe7f455e82b2428be8f2"} err="failed to get container status \"8fda896714711d9ce9687adc74a30b992983e810fbe4fe7f455e82b2428be8f2\": rpc error: code = NotFound desc = could not find container \"8fda896714711d9ce9687adc74a30b992983e810fbe4fe7f455e82b2428be8f2\": container with ID starting with 8fda896714711d9ce9687adc74a30b992983e810fbe4fe7f455e82b2428be8f2 not found: ID does not exist" Nov 24 18:11:56 crc kubenswrapper[4808]: I1124 18:11:56.357710 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99e39212-efe7-46b3-a03b-77d7110ade81" path="/var/lib/kubelet/pods/99e39212-efe7-46b3-a03b-77d7110ade81/volumes" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.189359 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nzj2t"] Nov 24 18:12:07 crc kubenswrapper[4808]: E1124 18:12:07.190639 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99e39212-efe7-46b3-a03b-77d7110ade81" containerName="extract-utilities" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.190660 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="99e39212-efe7-46b3-a03b-77d7110ade81" containerName="extract-utilities" Nov 24 18:12:07 crc kubenswrapper[4808]: E1124 18:12:07.190673 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99e39212-efe7-46b3-a03b-77d7110ade81" containerName="registry-server" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.190681 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="99e39212-efe7-46b3-a03b-77d7110ade81" containerName="registry-server" Nov 24 18:12:07 crc kubenswrapper[4808]: E1124 18:12:07.190734 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99e39212-efe7-46b3-a03b-77d7110ade81" containerName="extract-content" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.190749 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="99e39212-efe7-46b3-a03b-77d7110ade81" containerName="extract-content" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.190957 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="99e39212-efe7-46b3-a03b-77d7110ade81" containerName="registry-server" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.193128 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.197956 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nzj2t"] Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.297610 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/466e62f7-86c3-4ade-8abe-2f17dec8cbba-utilities\") pod \"certified-operators-nzj2t\" (UID: \"466e62f7-86c3-4ade-8abe-2f17dec8cbba\") " pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.297751 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/466e62f7-86c3-4ade-8abe-2f17dec8cbba-catalog-content\") pod \"certified-operators-nzj2t\" (UID: \"466e62f7-86c3-4ade-8abe-2f17dec8cbba\") " pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.297823 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfzk6\" (UniqueName: \"kubernetes.io/projected/466e62f7-86c3-4ade-8abe-2f17dec8cbba-kube-api-access-rfzk6\") pod \"certified-operators-nzj2t\" (UID: \"466e62f7-86c3-4ade-8abe-2f17dec8cbba\") " pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.400705 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/466e62f7-86c3-4ade-8abe-2f17dec8cbba-utilities\") pod \"certified-operators-nzj2t\" (UID: \"466e62f7-86c3-4ade-8abe-2f17dec8cbba\") " pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.400783 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/466e62f7-86c3-4ade-8abe-2f17dec8cbba-catalog-content\") pod \"certified-operators-nzj2t\" (UID: \"466e62f7-86c3-4ade-8abe-2f17dec8cbba\") " pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.400816 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfzk6\" (UniqueName: \"kubernetes.io/projected/466e62f7-86c3-4ade-8abe-2f17dec8cbba-kube-api-access-rfzk6\") pod \"certified-operators-nzj2t\" (UID: \"466e62f7-86c3-4ade-8abe-2f17dec8cbba\") " pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.401179 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/466e62f7-86c3-4ade-8abe-2f17dec8cbba-utilities\") pod \"certified-operators-nzj2t\" (UID: \"466e62f7-86c3-4ade-8abe-2f17dec8cbba\") " pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.401335 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/466e62f7-86c3-4ade-8abe-2f17dec8cbba-catalog-content\") pod \"certified-operators-nzj2t\" (UID: \"466e62f7-86c3-4ade-8abe-2f17dec8cbba\") " pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.431499 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfzk6\" (UniqueName: \"kubernetes.io/projected/466e62f7-86c3-4ade-8abe-2f17dec8cbba-kube-api-access-rfzk6\") pod \"certified-operators-nzj2t\" (UID: \"466e62f7-86c3-4ade-8abe-2f17dec8cbba\") " pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:07 crc kubenswrapper[4808]: I1124 18:12:07.530526 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:08 crc kubenswrapper[4808]: I1124 18:12:08.050574 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nzj2t"] Nov 24 18:12:08 crc kubenswrapper[4808]: I1124 18:12:08.922800 4808 generic.go:334] "Generic (PLEG): container finished" podID="466e62f7-86c3-4ade-8abe-2f17dec8cbba" containerID="fd9d0bbfd02674b3104bde28f477671b566bd1e781598f87ecdbd81395f98abc" exitCode=0 Nov 24 18:12:08 crc kubenswrapper[4808]: I1124 18:12:08.922863 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzj2t" event={"ID":"466e62f7-86c3-4ade-8abe-2f17dec8cbba","Type":"ContainerDied","Data":"fd9d0bbfd02674b3104bde28f477671b566bd1e781598f87ecdbd81395f98abc"} Nov 24 18:12:08 crc kubenswrapper[4808]: I1124 18:12:08.923210 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzj2t" event={"ID":"466e62f7-86c3-4ade-8abe-2f17dec8cbba","Type":"ContainerStarted","Data":"cf78e99e9219bb6e71c5ec9e18c5c3def9258444b7d0517941e680ea7831f618"} Nov 24 18:12:09 crc kubenswrapper[4808]: I1124 18:12:09.935480 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzj2t" event={"ID":"466e62f7-86c3-4ade-8abe-2f17dec8cbba","Type":"ContainerStarted","Data":"8dc6d32ba7f63a8adc2bbdb5b03989856cbdbeca623a416c45f73fcdea602521"} Nov 24 18:12:10 crc kubenswrapper[4808]: I1124 18:12:10.952321 4808 generic.go:334] "Generic (PLEG): container finished" podID="466e62f7-86c3-4ade-8abe-2f17dec8cbba" containerID="8dc6d32ba7f63a8adc2bbdb5b03989856cbdbeca623a416c45f73fcdea602521" exitCode=0 Nov 24 18:12:10 crc kubenswrapper[4808]: I1124 18:12:10.952374 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzj2t" event={"ID":"466e62f7-86c3-4ade-8abe-2f17dec8cbba","Type":"ContainerDied","Data":"8dc6d32ba7f63a8adc2bbdb5b03989856cbdbeca623a416c45f73fcdea602521"} Nov 24 18:12:11 crc kubenswrapper[4808]: I1124 18:12:11.985380 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzj2t" event={"ID":"466e62f7-86c3-4ade-8abe-2f17dec8cbba","Type":"ContainerStarted","Data":"50e4deac31e9d7e229b2cf28d5833260a634840459d599e91b1c62413eb573c0"} Nov 24 18:12:17 crc kubenswrapper[4808]: I1124 18:12:17.530721 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:17 crc kubenswrapper[4808]: I1124 18:12:17.531251 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:17 crc kubenswrapper[4808]: I1124 18:12:17.587576 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:17 crc kubenswrapper[4808]: I1124 18:12:17.611536 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nzj2t" podStartSLOduration=7.927694168 podStartE2EDuration="10.611516267s" podCreationTimestamp="2025-11-24 18:12:07 +0000 UTC" firstStartedPulling="2025-11-24 18:12:08.925832099 +0000 UTC m=+2721.523499941" lastFinishedPulling="2025-11-24 18:12:11.609654228 +0000 UTC m=+2724.207322040" observedRunningTime="2025-11-24 18:12:12.010316242 +0000 UTC m=+2724.607984044" watchObservedRunningTime="2025-11-24 18:12:17.611516267 +0000 UTC m=+2730.209184069" Nov 24 18:12:18 crc kubenswrapper[4808]: I1124 18:12:18.094436 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:18 crc kubenswrapper[4808]: I1124 18:12:18.145493 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nzj2t"] Nov 24 18:12:20 crc kubenswrapper[4808]: I1124 18:12:20.068459 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nzj2t" podUID="466e62f7-86c3-4ade-8abe-2f17dec8cbba" containerName="registry-server" containerID="cri-o://50e4deac31e9d7e229b2cf28d5833260a634840459d599e91b1c62413eb573c0" gracePeriod=2 Nov 24 18:12:20 crc kubenswrapper[4808]: I1124 18:12:20.623153 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:20 crc kubenswrapper[4808]: I1124 18:12:20.760288 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/466e62f7-86c3-4ade-8abe-2f17dec8cbba-catalog-content\") pod \"466e62f7-86c3-4ade-8abe-2f17dec8cbba\" (UID: \"466e62f7-86c3-4ade-8abe-2f17dec8cbba\") " Nov 24 18:12:20 crc kubenswrapper[4808]: I1124 18:12:20.760430 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfzk6\" (UniqueName: \"kubernetes.io/projected/466e62f7-86c3-4ade-8abe-2f17dec8cbba-kube-api-access-rfzk6\") pod \"466e62f7-86c3-4ade-8abe-2f17dec8cbba\" (UID: \"466e62f7-86c3-4ade-8abe-2f17dec8cbba\") " Nov 24 18:12:20 crc kubenswrapper[4808]: I1124 18:12:20.760531 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/466e62f7-86c3-4ade-8abe-2f17dec8cbba-utilities\") pod \"466e62f7-86c3-4ade-8abe-2f17dec8cbba\" (UID: \"466e62f7-86c3-4ade-8abe-2f17dec8cbba\") " Nov 24 18:12:20 crc kubenswrapper[4808]: I1124 18:12:20.761757 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/466e62f7-86c3-4ade-8abe-2f17dec8cbba-utilities" (OuterVolumeSpecName: "utilities") pod "466e62f7-86c3-4ade-8abe-2f17dec8cbba" (UID: "466e62f7-86c3-4ade-8abe-2f17dec8cbba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:12:20 crc kubenswrapper[4808]: I1124 18:12:20.771204 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/466e62f7-86c3-4ade-8abe-2f17dec8cbba-kube-api-access-rfzk6" (OuterVolumeSpecName: "kube-api-access-rfzk6") pod "466e62f7-86c3-4ade-8abe-2f17dec8cbba" (UID: "466e62f7-86c3-4ade-8abe-2f17dec8cbba"). InnerVolumeSpecName "kube-api-access-rfzk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:12:20 crc kubenswrapper[4808]: I1124 18:12:20.864241 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfzk6\" (UniqueName: \"kubernetes.io/projected/466e62f7-86c3-4ade-8abe-2f17dec8cbba-kube-api-access-rfzk6\") on node \"crc\" DevicePath \"\"" Nov 24 18:12:20 crc kubenswrapper[4808]: I1124 18:12:20.864491 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/466e62f7-86c3-4ade-8abe-2f17dec8cbba-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.068705 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/466e62f7-86c3-4ade-8abe-2f17dec8cbba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "466e62f7-86c3-4ade-8abe-2f17dec8cbba" (UID: "466e62f7-86c3-4ade-8abe-2f17dec8cbba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.092241 4808 generic.go:334] "Generic (PLEG): container finished" podID="466e62f7-86c3-4ade-8abe-2f17dec8cbba" containerID="50e4deac31e9d7e229b2cf28d5833260a634840459d599e91b1c62413eb573c0" exitCode=0 Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.092310 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzj2t" event={"ID":"466e62f7-86c3-4ade-8abe-2f17dec8cbba","Type":"ContainerDied","Data":"50e4deac31e9d7e229b2cf28d5833260a634840459d599e91b1c62413eb573c0"} Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.092352 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzj2t" event={"ID":"466e62f7-86c3-4ade-8abe-2f17dec8cbba","Type":"ContainerDied","Data":"cf78e99e9219bb6e71c5ec9e18c5c3def9258444b7d0517941e680ea7831f618"} Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.092383 4808 scope.go:117] "RemoveContainer" containerID="50e4deac31e9d7e229b2cf28d5833260a634840459d599e91b1c62413eb573c0" Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.092613 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nzj2t" Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.106933 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/466e62f7-86c3-4ade-8abe-2f17dec8cbba-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.154427 4808 scope.go:117] "RemoveContainer" containerID="8dc6d32ba7f63a8adc2bbdb5b03989856cbdbeca623a416c45f73fcdea602521" Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.160298 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nzj2t"] Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.184299 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nzj2t"] Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.189423 4808 scope.go:117] "RemoveContainer" containerID="fd9d0bbfd02674b3104bde28f477671b566bd1e781598f87ecdbd81395f98abc" Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.244706 4808 scope.go:117] "RemoveContainer" containerID="50e4deac31e9d7e229b2cf28d5833260a634840459d599e91b1c62413eb573c0" Nov 24 18:12:21 crc kubenswrapper[4808]: E1124 18:12:21.245198 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50e4deac31e9d7e229b2cf28d5833260a634840459d599e91b1c62413eb573c0\": container with ID starting with 50e4deac31e9d7e229b2cf28d5833260a634840459d599e91b1c62413eb573c0 not found: ID does not exist" containerID="50e4deac31e9d7e229b2cf28d5833260a634840459d599e91b1c62413eb573c0" Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.245240 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50e4deac31e9d7e229b2cf28d5833260a634840459d599e91b1c62413eb573c0"} err="failed to get container status \"50e4deac31e9d7e229b2cf28d5833260a634840459d599e91b1c62413eb573c0\": rpc error: code = NotFound desc = could not find container \"50e4deac31e9d7e229b2cf28d5833260a634840459d599e91b1c62413eb573c0\": container with ID starting with 50e4deac31e9d7e229b2cf28d5833260a634840459d599e91b1c62413eb573c0 not found: ID does not exist" Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.245269 4808 scope.go:117] "RemoveContainer" containerID="8dc6d32ba7f63a8adc2bbdb5b03989856cbdbeca623a416c45f73fcdea602521" Nov 24 18:12:21 crc kubenswrapper[4808]: E1124 18:12:21.245591 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dc6d32ba7f63a8adc2bbdb5b03989856cbdbeca623a416c45f73fcdea602521\": container with ID starting with 8dc6d32ba7f63a8adc2bbdb5b03989856cbdbeca623a416c45f73fcdea602521 not found: ID does not exist" containerID="8dc6d32ba7f63a8adc2bbdb5b03989856cbdbeca623a416c45f73fcdea602521" Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.245615 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dc6d32ba7f63a8adc2bbdb5b03989856cbdbeca623a416c45f73fcdea602521"} err="failed to get container status \"8dc6d32ba7f63a8adc2bbdb5b03989856cbdbeca623a416c45f73fcdea602521\": rpc error: code = NotFound desc = could not find container \"8dc6d32ba7f63a8adc2bbdb5b03989856cbdbeca623a416c45f73fcdea602521\": container with ID starting with 8dc6d32ba7f63a8adc2bbdb5b03989856cbdbeca623a416c45f73fcdea602521 not found: ID does not exist" Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.245628 4808 scope.go:117] "RemoveContainer" containerID="fd9d0bbfd02674b3104bde28f477671b566bd1e781598f87ecdbd81395f98abc" Nov 24 18:12:21 crc kubenswrapper[4808]: E1124 18:12:21.246205 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd9d0bbfd02674b3104bde28f477671b566bd1e781598f87ecdbd81395f98abc\": container with ID starting with fd9d0bbfd02674b3104bde28f477671b566bd1e781598f87ecdbd81395f98abc not found: ID does not exist" containerID="fd9d0bbfd02674b3104bde28f477671b566bd1e781598f87ecdbd81395f98abc" Nov 24 18:12:21 crc kubenswrapper[4808]: I1124 18:12:21.246283 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd9d0bbfd02674b3104bde28f477671b566bd1e781598f87ecdbd81395f98abc"} err="failed to get container status \"fd9d0bbfd02674b3104bde28f477671b566bd1e781598f87ecdbd81395f98abc\": rpc error: code = NotFound desc = could not find container \"fd9d0bbfd02674b3104bde28f477671b566bd1e781598f87ecdbd81395f98abc\": container with ID starting with fd9d0bbfd02674b3104bde28f477671b566bd1e781598f87ecdbd81395f98abc not found: ID does not exist" Nov 24 18:12:22 crc kubenswrapper[4808]: I1124 18:12:22.360783 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="466e62f7-86c3-4ade-8abe-2f17dec8cbba" path="/var/lib/kubelet/pods/466e62f7-86c3-4ade-8abe-2f17dec8cbba/volumes" Nov 24 18:13:06 crc kubenswrapper[4808]: I1124 18:13:06.523212 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:13:06 crc kubenswrapper[4808]: I1124 18:13:06.523843 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:13:36 crc kubenswrapper[4808]: I1124 18:13:36.523594 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:13:36 crc kubenswrapper[4808]: I1124 18:13:36.524361 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:14:06 crc kubenswrapper[4808]: I1124 18:14:06.522953 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:14:06 crc kubenswrapper[4808]: I1124 18:14:06.523425 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:14:06 crc kubenswrapper[4808]: I1124 18:14:06.523466 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 18:14:06 crc kubenswrapper[4808]: I1124 18:14:06.524167 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"40b4e634cbb4a00624c13f81bfdf705c343a7ea5c4f0bdc8a8675135a0436bfa"} pod="openshift-machine-config-operator/machine-config-daemon-ht87w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 18:14:06 crc kubenswrapper[4808]: I1124 18:14:06.524215 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" containerID="cri-o://40b4e634cbb4a00624c13f81bfdf705c343a7ea5c4f0bdc8a8675135a0436bfa" gracePeriod=600 Nov 24 18:14:07 crc kubenswrapper[4808]: I1124 18:14:07.217870 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerID="40b4e634cbb4a00624c13f81bfdf705c343a7ea5c4f0bdc8a8675135a0436bfa" exitCode=0 Nov 24 18:14:07 crc kubenswrapper[4808]: I1124 18:14:07.217946 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerDied","Data":"40b4e634cbb4a00624c13f81bfdf705c343a7ea5c4f0bdc8a8675135a0436bfa"} Nov 24 18:14:07 crc kubenswrapper[4808]: I1124 18:14:07.218236 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff"} Nov 24 18:14:07 crc kubenswrapper[4808]: I1124 18:14:07.218261 4808 scope.go:117] "RemoveContainer" containerID="493a9083db3b95ec39674839e467c6d2912e726af9531f5021110a996397162f" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.149652 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5"] Nov 24 18:15:00 crc kubenswrapper[4808]: E1124 18:15:00.150600 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="466e62f7-86c3-4ade-8abe-2f17dec8cbba" containerName="extract-utilities" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.150614 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="466e62f7-86c3-4ade-8abe-2f17dec8cbba" containerName="extract-utilities" Nov 24 18:15:00 crc kubenswrapper[4808]: E1124 18:15:00.150629 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="466e62f7-86c3-4ade-8abe-2f17dec8cbba" containerName="extract-content" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.150635 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="466e62f7-86c3-4ade-8abe-2f17dec8cbba" containerName="extract-content" Nov 24 18:15:00 crc kubenswrapper[4808]: E1124 18:15:00.150654 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="466e62f7-86c3-4ade-8abe-2f17dec8cbba" containerName="registry-server" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.150662 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="466e62f7-86c3-4ade-8abe-2f17dec8cbba" containerName="registry-server" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.150844 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="466e62f7-86c3-4ade-8abe-2f17dec8cbba" containerName="registry-server" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.151485 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.157401 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.157582 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.159430 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5"] Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.283569 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95gvz\" (UniqueName: \"kubernetes.io/projected/14ddd6f4-18be-44e8-8078-8a8647fdffc2-kube-api-access-95gvz\") pod \"collect-profiles-29400135-ngmg5\" (UID: \"14ddd6f4-18be-44e8-8078-8a8647fdffc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.283647 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/14ddd6f4-18be-44e8-8078-8a8647fdffc2-secret-volume\") pod \"collect-profiles-29400135-ngmg5\" (UID: \"14ddd6f4-18be-44e8-8078-8a8647fdffc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.283949 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/14ddd6f4-18be-44e8-8078-8a8647fdffc2-config-volume\") pod \"collect-profiles-29400135-ngmg5\" (UID: \"14ddd6f4-18be-44e8-8078-8a8647fdffc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.385644 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/14ddd6f4-18be-44e8-8078-8a8647fdffc2-config-volume\") pod \"collect-profiles-29400135-ngmg5\" (UID: \"14ddd6f4-18be-44e8-8078-8a8647fdffc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.385794 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95gvz\" (UniqueName: \"kubernetes.io/projected/14ddd6f4-18be-44e8-8078-8a8647fdffc2-kube-api-access-95gvz\") pod \"collect-profiles-29400135-ngmg5\" (UID: \"14ddd6f4-18be-44e8-8078-8a8647fdffc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.385843 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/14ddd6f4-18be-44e8-8078-8a8647fdffc2-secret-volume\") pod \"collect-profiles-29400135-ngmg5\" (UID: \"14ddd6f4-18be-44e8-8078-8a8647fdffc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.387362 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/14ddd6f4-18be-44e8-8078-8a8647fdffc2-config-volume\") pod \"collect-profiles-29400135-ngmg5\" (UID: \"14ddd6f4-18be-44e8-8078-8a8647fdffc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.394315 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/14ddd6f4-18be-44e8-8078-8a8647fdffc2-secret-volume\") pod \"collect-profiles-29400135-ngmg5\" (UID: \"14ddd6f4-18be-44e8-8078-8a8647fdffc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.403193 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95gvz\" (UniqueName: \"kubernetes.io/projected/14ddd6f4-18be-44e8-8078-8a8647fdffc2-kube-api-access-95gvz\") pod \"collect-profiles-29400135-ngmg5\" (UID: \"14ddd6f4-18be-44e8-8078-8a8647fdffc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.473875 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" Nov 24 18:15:00 crc kubenswrapper[4808]: I1124 18:15:00.944606 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5"] Nov 24 18:15:00 crc kubenswrapper[4808]: W1124 18:15:00.961405 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14ddd6f4_18be_44e8_8078_8a8647fdffc2.slice/crio-71ac79c2e5c8a11e371923dc732c188dcf9077ff88c1d753df46535de976a8d3 WatchSource:0}: Error finding container 71ac79c2e5c8a11e371923dc732c188dcf9077ff88c1d753df46535de976a8d3: Status 404 returned error can't find the container with id 71ac79c2e5c8a11e371923dc732c188dcf9077ff88c1d753df46535de976a8d3 Nov 24 18:15:01 crc kubenswrapper[4808]: I1124 18:15:01.766776 4808 generic.go:334] "Generic (PLEG): container finished" podID="14ddd6f4-18be-44e8-8078-8a8647fdffc2" containerID="b55671b54dab3701fdfe6dba35c43272f692867b61061c84b685e7df5465424d" exitCode=0 Nov 24 18:15:01 crc kubenswrapper[4808]: I1124 18:15:01.766930 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" event={"ID":"14ddd6f4-18be-44e8-8078-8a8647fdffc2","Type":"ContainerDied","Data":"b55671b54dab3701fdfe6dba35c43272f692867b61061c84b685e7df5465424d"} Nov 24 18:15:01 crc kubenswrapper[4808]: I1124 18:15:01.767190 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" event={"ID":"14ddd6f4-18be-44e8-8078-8a8647fdffc2","Type":"ContainerStarted","Data":"71ac79c2e5c8a11e371923dc732c188dcf9077ff88c1d753df46535de976a8d3"} Nov 24 18:15:03 crc kubenswrapper[4808]: I1124 18:15:03.088887 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" Nov 24 18:15:03 crc kubenswrapper[4808]: I1124 18:15:03.144267 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/14ddd6f4-18be-44e8-8078-8a8647fdffc2-config-volume\") pod \"14ddd6f4-18be-44e8-8078-8a8647fdffc2\" (UID: \"14ddd6f4-18be-44e8-8078-8a8647fdffc2\") " Nov 24 18:15:03 crc kubenswrapper[4808]: I1124 18:15:03.144361 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95gvz\" (UniqueName: \"kubernetes.io/projected/14ddd6f4-18be-44e8-8078-8a8647fdffc2-kube-api-access-95gvz\") pod \"14ddd6f4-18be-44e8-8078-8a8647fdffc2\" (UID: \"14ddd6f4-18be-44e8-8078-8a8647fdffc2\") " Nov 24 18:15:03 crc kubenswrapper[4808]: I1124 18:15:03.144438 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/14ddd6f4-18be-44e8-8078-8a8647fdffc2-secret-volume\") pod \"14ddd6f4-18be-44e8-8078-8a8647fdffc2\" (UID: \"14ddd6f4-18be-44e8-8078-8a8647fdffc2\") " Nov 24 18:15:03 crc kubenswrapper[4808]: I1124 18:15:03.145778 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14ddd6f4-18be-44e8-8078-8a8647fdffc2-config-volume" (OuterVolumeSpecName: "config-volume") pod "14ddd6f4-18be-44e8-8078-8a8647fdffc2" (UID: "14ddd6f4-18be-44e8-8078-8a8647fdffc2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 18:15:03 crc kubenswrapper[4808]: I1124 18:15:03.151585 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ddd6f4-18be-44e8-8078-8a8647fdffc2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "14ddd6f4-18be-44e8-8078-8a8647fdffc2" (UID: "14ddd6f4-18be-44e8-8078-8a8647fdffc2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:15:03 crc kubenswrapper[4808]: I1124 18:15:03.151873 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14ddd6f4-18be-44e8-8078-8a8647fdffc2-kube-api-access-95gvz" (OuterVolumeSpecName: "kube-api-access-95gvz") pod "14ddd6f4-18be-44e8-8078-8a8647fdffc2" (UID: "14ddd6f4-18be-44e8-8078-8a8647fdffc2"). InnerVolumeSpecName "kube-api-access-95gvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:15:03 crc kubenswrapper[4808]: I1124 18:15:03.246604 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/14ddd6f4-18be-44e8-8078-8a8647fdffc2-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 18:15:03 crc kubenswrapper[4808]: I1124 18:15:03.246655 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95gvz\" (UniqueName: \"kubernetes.io/projected/14ddd6f4-18be-44e8-8078-8a8647fdffc2-kube-api-access-95gvz\") on node \"crc\" DevicePath \"\"" Nov 24 18:15:03 crc kubenswrapper[4808]: I1124 18:15:03.246679 4808 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/14ddd6f4-18be-44e8-8078-8a8647fdffc2-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 18:15:03 crc kubenswrapper[4808]: I1124 18:15:03.785820 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" event={"ID":"14ddd6f4-18be-44e8-8078-8a8647fdffc2","Type":"ContainerDied","Data":"71ac79c2e5c8a11e371923dc732c188dcf9077ff88c1d753df46535de976a8d3"} Nov 24 18:15:03 crc kubenswrapper[4808]: I1124 18:15:03.786312 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71ac79c2e5c8a11e371923dc732c188dcf9077ff88c1d753df46535de976a8d3" Nov 24 18:15:03 crc kubenswrapper[4808]: I1124 18:15:03.786068 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400135-ngmg5" Nov 24 18:15:04 crc kubenswrapper[4808]: I1124 18:15:04.193656 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz"] Nov 24 18:15:04 crc kubenswrapper[4808]: I1124 18:15:04.202817 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400090-d86cz"] Nov 24 18:15:04 crc kubenswrapper[4808]: I1124 18:15:04.364688 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fee761ab-3ef8-4ee5-9bea-59395abe3fbc" path="/var/lib/kubelet/pods/fee761ab-3ef8-4ee5-9bea-59395abe3fbc/volumes" Nov 24 18:15:17 crc kubenswrapper[4808]: I1124 18:15:17.894169 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-78vlg"] Nov 24 18:15:17 crc kubenswrapper[4808]: E1124 18:15:17.895312 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ddd6f4-18be-44e8-8078-8a8647fdffc2" containerName="collect-profiles" Nov 24 18:15:17 crc kubenswrapper[4808]: I1124 18:15:17.895328 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ddd6f4-18be-44e8-8078-8a8647fdffc2" containerName="collect-profiles" Nov 24 18:15:17 crc kubenswrapper[4808]: I1124 18:15:17.895563 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ddd6f4-18be-44e8-8078-8a8647fdffc2" containerName="collect-profiles" Nov 24 18:15:17 crc kubenswrapper[4808]: I1124 18:15:17.896909 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:17 crc kubenswrapper[4808]: I1124 18:15:17.923968 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-78vlg"] Nov 24 18:15:17 crc kubenswrapper[4808]: I1124 18:15:17.993068 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5xwb\" (UniqueName: \"kubernetes.io/projected/5d84981a-9e5a-4d78-b46c-911c799331e6-kube-api-access-n5xwb\") pod \"redhat-marketplace-78vlg\" (UID: \"5d84981a-9e5a-4d78-b46c-911c799331e6\") " pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:17 crc kubenswrapper[4808]: I1124 18:15:17.993148 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d84981a-9e5a-4d78-b46c-911c799331e6-utilities\") pod \"redhat-marketplace-78vlg\" (UID: \"5d84981a-9e5a-4d78-b46c-911c799331e6\") " pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:17 crc kubenswrapper[4808]: I1124 18:15:17.993346 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d84981a-9e5a-4d78-b46c-911c799331e6-catalog-content\") pod \"redhat-marketplace-78vlg\" (UID: \"5d84981a-9e5a-4d78-b46c-911c799331e6\") " pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:18 crc kubenswrapper[4808]: I1124 18:15:18.094834 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5xwb\" (UniqueName: \"kubernetes.io/projected/5d84981a-9e5a-4d78-b46c-911c799331e6-kube-api-access-n5xwb\") pod \"redhat-marketplace-78vlg\" (UID: \"5d84981a-9e5a-4d78-b46c-911c799331e6\") " pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:18 crc kubenswrapper[4808]: I1124 18:15:18.094906 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d84981a-9e5a-4d78-b46c-911c799331e6-utilities\") pod \"redhat-marketplace-78vlg\" (UID: \"5d84981a-9e5a-4d78-b46c-911c799331e6\") " pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:18 crc kubenswrapper[4808]: I1124 18:15:18.095007 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d84981a-9e5a-4d78-b46c-911c799331e6-catalog-content\") pod \"redhat-marketplace-78vlg\" (UID: \"5d84981a-9e5a-4d78-b46c-911c799331e6\") " pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:18 crc kubenswrapper[4808]: I1124 18:15:18.095639 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d84981a-9e5a-4d78-b46c-911c799331e6-catalog-content\") pod \"redhat-marketplace-78vlg\" (UID: \"5d84981a-9e5a-4d78-b46c-911c799331e6\") " pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:18 crc kubenswrapper[4808]: I1124 18:15:18.095639 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d84981a-9e5a-4d78-b46c-911c799331e6-utilities\") pod \"redhat-marketplace-78vlg\" (UID: \"5d84981a-9e5a-4d78-b46c-911c799331e6\") " pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:18 crc kubenswrapper[4808]: I1124 18:15:18.117250 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5xwb\" (UniqueName: \"kubernetes.io/projected/5d84981a-9e5a-4d78-b46c-911c799331e6-kube-api-access-n5xwb\") pod \"redhat-marketplace-78vlg\" (UID: \"5d84981a-9e5a-4d78-b46c-911c799331e6\") " pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:18 crc kubenswrapper[4808]: I1124 18:15:18.221238 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:18 crc kubenswrapper[4808]: W1124 18:15:18.720387 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d84981a_9e5a_4d78_b46c_911c799331e6.slice/crio-fe613750a19f207045dd5bce256192624f440b000b19e267c321bbc66493b9a7 WatchSource:0}: Error finding container fe613750a19f207045dd5bce256192624f440b000b19e267c321bbc66493b9a7: Status 404 returned error can't find the container with id fe613750a19f207045dd5bce256192624f440b000b19e267c321bbc66493b9a7 Nov 24 18:15:18 crc kubenswrapper[4808]: I1124 18:15:18.723136 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-78vlg"] Nov 24 18:15:18 crc kubenswrapper[4808]: I1124 18:15:18.934390 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78vlg" event={"ID":"5d84981a-9e5a-4d78-b46c-911c799331e6","Type":"ContainerStarted","Data":"fe613750a19f207045dd5bce256192624f440b000b19e267c321bbc66493b9a7"} Nov 24 18:15:19 crc kubenswrapper[4808]: I1124 18:15:19.948539 4808 generic.go:334] "Generic (PLEG): container finished" podID="5d84981a-9e5a-4d78-b46c-911c799331e6" containerID="898de22e1234224ce3adcf38a2c72b3db458e782dfe94d63a973b96d33a05ed9" exitCode=0 Nov 24 18:15:19 crc kubenswrapper[4808]: I1124 18:15:19.948738 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78vlg" event={"ID":"5d84981a-9e5a-4d78-b46c-911c799331e6","Type":"ContainerDied","Data":"898de22e1234224ce3adcf38a2c72b3db458e782dfe94d63a973b96d33a05ed9"} Nov 24 18:15:20 crc kubenswrapper[4808]: I1124 18:15:20.958321 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78vlg" event={"ID":"5d84981a-9e5a-4d78-b46c-911c799331e6","Type":"ContainerStarted","Data":"34dc8b0071e05c0acfcdc8936f6c8c5e7603feca5eb2cffa818fa9894861d11b"} Nov 24 18:15:21 crc kubenswrapper[4808]: I1124 18:15:21.974723 4808 generic.go:334] "Generic (PLEG): container finished" podID="5d84981a-9e5a-4d78-b46c-911c799331e6" containerID="34dc8b0071e05c0acfcdc8936f6c8c5e7603feca5eb2cffa818fa9894861d11b" exitCode=0 Nov 24 18:15:21 crc kubenswrapper[4808]: I1124 18:15:21.974779 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78vlg" event={"ID":"5d84981a-9e5a-4d78-b46c-911c799331e6","Type":"ContainerDied","Data":"34dc8b0071e05c0acfcdc8936f6c8c5e7603feca5eb2cffa818fa9894861d11b"} Nov 24 18:15:22 crc kubenswrapper[4808]: I1124 18:15:22.990220 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78vlg" event={"ID":"5d84981a-9e5a-4d78-b46c-911c799331e6","Type":"ContainerStarted","Data":"5eb968e2754bb7cca676048932fad775b2e72b5065f0738c8b7587404d82f968"} Nov 24 18:15:23 crc kubenswrapper[4808]: I1124 18:15:23.016271 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-78vlg" podStartSLOduration=3.574477583 podStartE2EDuration="6.016249683s" podCreationTimestamp="2025-11-24 18:15:17 +0000 UTC" firstStartedPulling="2025-11-24 18:15:19.950318633 +0000 UTC m=+2912.547986435" lastFinishedPulling="2025-11-24 18:15:22.392090743 +0000 UTC m=+2914.989758535" observedRunningTime="2025-11-24 18:15:23.007055904 +0000 UTC m=+2915.604723716" watchObservedRunningTime="2025-11-24 18:15:23.016249683 +0000 UTC m=+2915.613917495" Nov 24 18:15:28 crc kubenswrapper[4808]: I1124 18:15:28.221640 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:28 crc kubenswrapper[4808]: I1124 18:15:28.222482 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:28 crc kubenswrapper[4808]: I1124 18:15:28.275857 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:29 crc kubenswrapper[4808]: I1124 18:15:29.111528 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:29 crc kubenswrapper[4808]: I1124 18:15:29.164154 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-78vlg"] Nov 24 18:15:31 crc kubenswrapper[4808]: I1124 18:15:31.068716 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-78vlg" podUID="5d84981a-9e5a-4d78-b46c-911c799331e6" containerName="registry-server" containerID="cri-o://5eb968e2754bb7cca676048932fad775b2e72b5065f0738c8b7587404d82f968" gracePeriod=2 Nov 24 18:15:31 crc kubenswrapper[4808]: I1124 18:15:31.588054 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:31 crc kubenswrapper[4808]: I1124 18:15:31.667469 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d84981a-9e5a-4d78-b46c-911c799331e6-utilities\") pod \"5d84981a-9e5a-4d78-b46c-911c799331e6\" (UID: \"5d84981a-9e5a-4d78-b46c-911c799331e6\") " Nov 24 18:15:31 crc kubenswrapper[4808]: I1124 18:15:31.667693 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5xwb\" (UniqueName: \"kubernetes.io/projected/5d84981a-9e5a-4d78-b46c-911c799331e6-kube-api-access-n5xwb\") pod \"5d84981a-9e5a-4d78-b46c-911c799331e6\" (UID: \"5d84981a-9e5a-4d78-b46c-911c799331e6\") " Nov 24 18:15:31 crc kubenswrapper[4808]: I1124 18:15:31.668444 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d84981a-9e5a-4d78-b46c-911c799331e6-utilities" (OuterVolumeSpecName: "utilities") pod "5d84981a-9e5a-4d78-b46c-911c799331e6" (UID: "5d84981a-9e5a-4d78-b46c-911c799331e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:15:31 crc kubenswrapper[4808]: I1124 18:15:31.669375 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d84981a-9e5a-4d78-b46c-911c799331e6-catalog-content\") pod \"5d84981a-9e5a-4d78-b46c-911c799331e6\" (UID: \"5d84981a-9e5a-4d78-b46c-911c799331e6\") " Nov 24 18:15:31 crc kubenswrapper[4808]: I1124 18:15:31.671706 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d84981a-9e5a-4d78-b46c-911c799331e6-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:15:31 crc kubenswrapper[4808]: I1124 18:15:31.674001 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d84981a-9e5a-4d78-b46c-911c799331e6-kube-api-access-n5xwb" (OuterVolumeSpecName: "kube-api-access-n5xwb") pod "5d84981a-9e5a-4d78-b46c-911c799331e6" (UID: "5d84981a-9e5a-4d78-b46c-911c799331e6"). InnerVolumeSpecName "kube-api-access-n5xwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:15:31 crc kubenswrapper[4808]: I1124 18:15:31.686882 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d84981a-9e5a-4d78-b46c-911c799331e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5d84981a-9e5a-4d78-b46c-911c799331e6" (UID: "5d84981a-9e5a-4d78-b46c-911c799331e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:15:31 crc kubenswrapper[4808]: I1124 18:15:31.773574 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5xwb\" (UniqueName: \"kubernetes.io/projected/5d84981a-9e5a-4d78-b46c-911c799331e6-kube-api-access-n5xwb\") on node \"crc\" DevicePath \"\"" Nov 24 18:15:31 crc kubenswrapper[4808]: I1124 18:15:31.773606 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d84981a-9e5a-4d78-b46c-911c799331e6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.087060 4808 generic.go:334] "Generic (PLEG): container finished" podID="5d84981a-9e5a-4d78-b46c-911c799331e6" containerID="5eb968e2754bb7cca676048932fad775b2e72b5065f0738c8b7587404d82f968" exitCode=0 Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.087107 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78vlg" event={"ID":"5d84981a-9e5a-4d78-b46c-911c799331e6","Type":"ContainerDied","Data":"5eb968e2754bb7cca676048932fad775b2e72b5065f0738c8b7587404d82f968"} Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.087141 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78vlg" event={"ID":"5d84981a-9e5a-4d78-b46c-911c799331e6","Type":"ContainerDied","Data":"fe613750a19f207045dd5bce256192624f440b000b19e267c321bbc66493b9a7"} Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.087169 4808 scope.go:117] "RemoveContainer" containerID="5eb968e2754bb7cca676048932fad775b2e72b5065f0738c8b7587404d82f968" Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.087235 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78vlg" Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.128885 4808 scope.go:117] "RemoveContainer" containerID="34dc8b0071e05c0acfcdc8936f6c8c5e7603feca5eb2cffa818fa9894861d11b" Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.147504 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-78vlg"] Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.156587 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-78vlg"] Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.166821 4808 scope.go:117] "RemoveContainer" containerID="898de22e1234224ce3adcf38a2c72b3db458e782dfe94d63a973b96d33a05ed9" Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.226215 4808 scope.go:117] "RemoveContainer" containerID="5eb968e2754bb7cca676048932fad775b2e72b5065f0738c8b7587404d82f968" Nov 24 18:15:32 crc kubenswrapper[4808]: E1124 18:15:32.230460 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eb968e2754bb7cca676048932fad775b2e72b5065f0738c8b7587404d82f968\": container with ID starting with 5eb968e2754bb7cca676048932fad775b2e72b5065f0738c8b7587404d82f968 not found: ID does not exist" containerID="5eb968e2754bb7cca676048932fad775b2e72b5065f0738c8b7587404d82f968" Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.230586 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eb968e2754bb7cca676048932fad775b2e72b5065f0738c8b7587404d82f968"} err="failed to get container status \"5eb968e2754bb7cca676048932fad775b2e72b5065f0738c8b7587404d82f968\": rpc error: code = NotFound desc = could not find container \"5eb968e2754bb7cca676048932fad775b2e72b5065f0738c8b7587404d82f968\": container with ID starting with 5eb968e2754bb7cca676048932fad775b2e72b5065f0738c8b7587404d82f968 not found: ID does not exist" Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.230660 4808 scope.go:117] "RemoveContainer" containerID="34dc8b0071e05c0acfcdc8936f6c8c5e7603feca5eb2cffa818fa9894861d11b" Nov 24 18:15:32 crc kubenswrapper[4808]: E1124 18:15:32.231667 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34dc8b0071e05c0acfcdc8936f6c8c5e7603feca5eb2cffa818fa9894861d11b\": container with ID starting with 34dc8b0071e05c0acfcdc8936f6c8c5e7603feca5eb2cffa818fa9894861d11b not found: ID does not exist" containerID="34dc8b0071e05c0acfcdc8936f6c8c5e7603feca5eb2cffa818fa9894861d11b" Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.231933 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34dc8b0071e05c0acfcdc8936f6c8c5e7603feca5eb2cffa818fa9894861d11b"} err="failed to get container status \"34dc8b0071e05c0acfcdc8936f6c8c5e7603feca5eb2cffa818fa9894861d11b\": rpc error: code = NotFound desc = could not find container \"34dc8b0071e05c0acfcdc8936f6c8c5e7603feca5eb2cffa818fa9894861d11b\": container with ID starting with 34dc8b0071e05c0acfcdc8936f6c8c5e7603feca5eb2cffa818fa9894861d11b not found: ID does not exist" Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.231974 4808 scope.go:117] "RemoveContainer" containerID="898de22e1234224ce3adcf38a2c72b3db458e782dfe94d63a973b96d33a05ed9" Nov 24 18:15:32 crc kubenswrapper[4808]: E1124 18:15:32.232493 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"898de22e1234224ce3adcf38a2c72b3db458e782dfe94d63a973b96d33a05ed9\": container with ID starting with 898de22e1234224ce3adcf38a2c72b3db458e782dfe94d63a973b96d33a05ed9 not found: ID does not exist" containerID="898de22e1234224ce3adcf38a2c72b3db458e782dfe94d63a973b96d33a05ed9" Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.232550 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"898de22e1234224ce3adcf38a2c72b3db458e782dfe94d63a973b96d33a05ed9"} err="failed to get container status \"898de22e1234224ce3adcf38a2c72b3db458e782dfe94d63a973b96d33a05ed9\": rpc error: code = NotFound desc = could not find container \"898de22e1234224ce3adcf38a2c72b3db458e782dfe94d63a973b96d33a05ed9\": container with ID starting with 898de22e1234224ce3adcf38a2c72b3db458e782dfe94d63a973b96d33a05ed9 not found: ID does not exist" Nov 24 18:15:32 crc kubenswrapper[4808]: I1124 18:15:32.357549 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d84981a-9e5a-4d78-b46c-911c799331e6" path="/var/lib/kubelet/pods/5d84981a-9e5a-4d78-b46c-911c799331e6/volumes" Nov 24 18:15:54 crc kubenswrapper[4808]: I1124 18:15:54.798695 4808 scope.go:117] "RemoveContainer" containerID="207281a72bfd1685b869102173cb9465159e31d100b42e88c9039e7273c86b88" Nov 24 18:16:06 crc kubenswrapper[4808]: I1124 18:16:06.523099 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:16:06 crc kubenswrapper[4808]: I1124 18:16:06.523804 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:16:36 crc kubenswrapper[4808]: I1124 18:16:36.523186 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:16:36 crc kubenswrapper[4808]: I1124 18:16:36.523948 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:17:06 crc kubenswrapper[4808]: I1124 18:17:06.523193 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:17:06 crc kubenswrapper[4808]: I1124 18:17:06.523803 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:17:06 crc kubenswrapper[4808]: I1124 18:17:06.523867 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 18:17:06 crc kubenswrapper[4808]: I1124 18:17:06.524906 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff"} pod="openshift-machine-config-operator/machine-config-daemon-ht87w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 18:17:06 crc kubenswrapper[4808]: I1124 18:17:06.525007 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" containerID="cri-o://b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" gracePeriod=600 Nov 24 18:17:06 crc kubenswrapper[4808]: E1124 18:17:06.677964 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:17:07 crc kubenswrapper[4808]: I1124 18:17:07.090914 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" exitCode=0 Nov 24 18:17:07 crc kubenswrapper[4808]: I1124 18:17:07.090962 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerDied","Data":"b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff"} Nov 24 18:17:07 crc kubenswrapper[4808]: I1124 18:17:07.090999 4808 scope.go:117] "RemoveContainer" containerID="40b4e634cbb4a00624c13f81bfdf705c343a7ea5c4f0bdc8a8675135a0436bfa" Nov 24 18:17:07 crc kubenswrapper[4808]: I1124 18:17:07.091693 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:17:07 crc kubenswrapper[4808]: E1124 18:17:07.092361 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:17:21 crc kubenswrapper[4808]: I1124 18:17:21.347392 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:17:21 crc kubenswrapper[4808]: E1124 18:17:21.348428 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:17:32 crc kubenswrapper[4808]: I1124 18:17:32.348276 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:17:32 crc kubenswrapper[4808]: E1124 18:17:32.349469 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:17:43 crc kubenswrapper[4808]: I1124 18:17:43.347912 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:17:43 crc kubenswrapper[4808]: E1124 18:17:43.348646 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:17:54 crc kubenswrapper[4808]: I1124 18:17:54.348226 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:17:54 crc kubenswrapper[4808]: E1124 18:17:54.349379 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:18:09 crc kubenswrapper[4808]: I1124 18:18:09.347753 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:18:09 crc kubenswrapper[4808]: E1124 18:18:09.349191 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:18:22 crc kubenswrapper[4808]: I1124 18:18:22.348792 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:18:22 crc kubenswrapper[4808]: E1124 18:18:22.350519 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:18:36 crc kubenswrapper[4808]: I1124 18:18:36.348176 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:18:36 crc kubenswrapper[4808]: E1124 18:18:36.349347 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:18:51 crc kubenswrapper[4808]: I1124 18:18:51.349540 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:18:51 crc kubenswrapper[4808]: E1124 18:18:51.353418 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:19:04 crc kubenswrapper[4808]: I1124 18:19:04.348208 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:19:04 crc kubenswrapper[4808]: E1124 18:19:04.349001 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:19:18 crc kubenswrapper[4808]: I1124 18:19:18.354047 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:19:18 crc kubenswrapper[4808]: E1124 18:19:18.354635 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:19:31 crc kubenswrapper[4808]: I1124 18:19:31.347959 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:19:31 crc kubenswrapper[4808]: E1124 18:19:31.348834 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:19:43 crc kubenswrapper[4808]: I1124 18:19:43.347459 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:19:43 crc kubenswrapper[4808]: E1124 18:19:43.348155 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:19:58 crc kubenswrapper[4808]: I1124 18:19:58.355318 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:19:58 crc kubenswrapper[4808]: E1124 18:19:58.356372 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:20:10 crc kubenswrapper[4808]: I1124 18:20:10.349011 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:20:10 crc kubenswrapper[4808]: E1124 18:20:10.350353 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:20:22 crc kubenswrapper[4808]: I1124 18:20:22.348182 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:20:22 crc kubenswrapper[4808]: E1124 18:20:22.349501 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:20:36 crc kubenswrapper[4808]: I1124 18:20:36.347058 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:20:36 crc kubenswrapper[4808]: E1124 18:20:36.348124 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:20:37 crc kubenswrapper[4808]: I1124 18:20:37.663951 4808 generic.go:334] "Generic (PLEG): container finished" podID="d014b1af-f795-43e4-af00-0a152d0cb945" containerID="72d4294f444f4381cf5070cb811ee1e5eb4f4c6b3578bdf87b2a75cd3a203826" exitCode=0 Nov 24 18:20:37 crc kubenswrapper[4808]: I1124 18:20:37.664085 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d014b1af-f795-43e4-af00-0a152d0cb945","Type":"ContainerDied","Data":"72d4294f444f4381cf5070cb811ee1e5eb4f4c6b3578bdf87b2a75cd3a203826"} Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.154363 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.314658 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d014b1af-f795-43e4-af00-0a152d0cb945-test-operator-ephemeral-temporary\") pod \"d014b1af-f795-43e4-af00-0a152d0cb945\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.314697 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-openstack-config-secret\") pod \"d014b1af-f795-43e4-af00-0a152d0cb945\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.314765 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-ca-certs\") pod \"d014b1af-f795-43e4-af00-0a152d0cb945\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.315533 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d014b1af-f795-43e4-af00-0a152d0cb945-config-data\") pod \"d014b1af-f795-43e4-af00-0a152d0cb945\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.315620 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g77gz\" (UniqueName: \"kubernetes.io/projected/d014b1af-f795-43e4-af00-0a152d0cb945-kube-api-access-g77gz\") pod \"d014b1af-f795-43e4-af00-0a152d0cb945\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.315684 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d014b1af-f795-43e4-af00-0a152d0cb945-openstack-config\") pod \"d014b1af-f795-43e4-af00-0a152d0cb945\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.315738 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d014b1af-f795-43e4-af00-0a152d0cb945-test-operator-ephemeral-workdir\") pod \"d014b1af-f795-43e4-af00-0a152d0cb945\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.315798 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"d014b1af-f795-43e4-af00-0a152d0cb945\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.315890 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-ssh-key\") pod \"d014b1af-f795-43e4-af00-0a152d0cb945\" (UID: \"d014b1af-f795-43e4-af00-0a152d0cb945\") " Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.316409 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d014b1af-f795-43e4-af00-0a152d0cb945-config-data" (OuterVolumeSpecName: "config-data") pod "d014b1af-f795-43e4-af00-0a152d0cb945" (UID: "d014b1af-f795-43e4-af00-0a152d0cb945"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.316866 4808 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d014b1af-f795-43e4-af00-0a152d0cb945-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.316898 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d014b1af-f795-43e4-af00-0a152d0cb945-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "d014b1af-f795-43e4-af00-0a152d0cb945" (UID: "d014b1af-f795-43e4-af00-0a152d0cb945"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.320206 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "test-operator-logs") pod "d014b1af-f795-43e4-af00-0a152d0cb945" (UID: "d014b1af-f795-43e4-af00-0a152d0cb945"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.320305 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d014b1af-f795-43e4-af00-0a152d0cb945-kube-api-access-g77gz" (OuterVolumeSpecName: "kube-api-access-g77gz") pod "d014b1af-f795-43e4-af00-0a152d0cb945" (UID: "d014b1af-f795-43e4-af00-0a152d0cb945"). InnerVolumeSpecName "kube-api-access-g77gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.330741 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d014b1af-f795-43e4-af00-0a152d0cb945-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "d014b1af-f795-43e4-af00-0a152d0cb945" (UID: "d014b1af-f795-43e4-af00-0a152d0cb945"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.340478 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "d014b1af-f795-43e4-af00-0a152d0cb945" (UID: "d014b1af-f795-43e4-af00-0a152d0cb945"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.341721 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d014b1af-f795-43e4-af00-0a152d0cb945" (UID: "d014b1af-f795-43e4-af00-0a152d0cb945"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.349489 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d014b1af-f795-43e4-af00-0a152d0cb945" (UID: "d014b1af-f795-43e4-af00-0a152d0cb945"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.376939 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d014b1af-f795-43e4-af00-0a152d0cb945-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "d014b1af-f795-43e4-af00-0a152d0cb945" (UID: "d014b1af-f795-43e4-af00-0a152d0cb945"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.417879 4808 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.417925 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g77gz\" (UniqueName: \"kubernetes.io/projected/d014b1af-f795-43e4-af00-0a152d0cb945-kube-api-access-g77gz\") on node \"crc\" DevicePath \"\"" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.417939 4808 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d014b1af-f795-43e4-af00-0a152d0cb945-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.417951 4808 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d014b1af-f795-43e4-af00-0a152d0cb945-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.417976 4808 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.417992 4808 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.418004 4808 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d014b1af-f795-43e4-af00-0a152d0cb945-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.418035 4808 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d014b1af-f795-43e4-af00-0a152d0cb945-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.435905 4808 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.520593 4808 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.692721 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d014b1af-f795-43e4-af00-0a152d0cb945","Type":"ContainerDied","Data":"e9dca01a08450a29b1101326a77535ac2472c681db1fe0acd2eadfbe47bf3853"} Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.692780 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9dca01a08450a29b1101326a77535ac2472c681db1fe0acd2eadfbe47bf3853" Nov 24 18:20:39 crc kubenswrapper[4808]: I1124 18:20:39.692830 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.594362 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 18:20:44 crc kubenswrapper[4808]: E1124 18:20:44.596155 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d84981a-9e5a-4d78-b46c-911c799331e6" containerName="extract-utilities" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.596184 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d84981a-9e5a-4d78-b46c-911c799331e6" containerName="extract-utilities" Nov 24 18:20:44 crc kubenswrapper[4808]: E1124 18:20:44.596370 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d014b1af-f795-43e4-af00-0a152d0cb945" containerName="tempest-tests-tempest-tests-runner" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.596391 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d014b1af-f795-43e4-af00-0a152d0cb945" containerName="tempest-tests-tempest-tests-runner" Nov 24 18:20:44 crc kubenswrapper[4808]: E1124 18:20:44.596428 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d84981a-9e5a-4d78-b46c-911c799331e6" containerName="registry-server" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.596442 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d84981a-9e5a-4d78-b46c-911c799331e6" containerName="registry-server" Nov 24 18:20:44 crc kubenswrapper[4808]: E1124 18:20:44.596486 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d84981a-9e5a-4d78-b46c-911c799331e6" containerName="extract-content" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.596540 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d84981a-9e5a-4d78-b46c-911c799331e6" containerName="extract-content" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.596973 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d84981a-9e5a-4d78-b46c-911c799331e6" containerName="registry-server" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.597011 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="d014b1af-f795-43e4-af00-0a152d0cb945" containerName="tempest-tests-tempest-tests-runner" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.598562 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.603010 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-xhqjb" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.606227 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.727896 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdxd8\" (UniqueName: \"kubernetes.io/projected/5c95e77e-1080-4ca8-96d8-8b34632303ea-kube-api-access-hdxd8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5c95e77e-1080-4ca8-96d8-8b34632303ea\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.727938 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5c95e77e-1080-4ca8-96d8-8b34632303ea\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.830200 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdxd8\" (UniqueName: \"kubernetes.io/projected/5c95e77e-1080-4ca8-96d8-8b34632303ea-kube-api-access-hdxd8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5c95e77e-1080-4ca8-96d8-8b34632303ea\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.830266 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5c95e77e-1080-4ca8-96d8-8b34632303ea\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.831138 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5c95e77e-1080-4ca8-96d8-8b34632303ea\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.857797 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdxd8\" (UniqueName: \"kubernetes.io/projected/5c95e77e-1080-4ca8-96d8-8b34632303ea-kube-api-access-hdxd8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5c95e77e-1080-4ca8-96d8-8b34632303ea\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.888322 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5c95e77e-1080-4ca8-96d8-8b34632303ea\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 18:20:44 crc kubenswrapper[4808]: I1124 18:20:44.947728 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 18:20:45 crc kubenswrapper[4808]: I1124 18:20:45.488101 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 18:20:45 crc kubenswrapper[4808]: I1124 18:20:45.493213 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 18:20:45 crc kubenswrapper[4808]: I1124 18:20:45.761428 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"5c95e77e-1080-4ca8-96d8-8b34632303ea","Type":"ContainerStarted","Data":"38a4415f162d01ebc8afab75e0e343285b1c111acfcf48efa78967487dad41df"} Nov 24 18:20:47 crc kubenswrapper[4808]: I1124 18:20:47.791462 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"5c95e77e-1080-4ca8-96d8-8b34632303ea","Type":"ContainerStarted","Data":"32d987b1119aa2df45dccaf8f858ec0d464f6a55cabb502c993901edebfb12a5"} Nov 24 18:20:47 crc kubenswrapper[4808]: I1124 18:20:47.810770 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.354937789 podStartE2EDuration="3.810740039s" podCreationTimestamp="2025-11-24 18:20:44 +0000 UTC" firstStartedPulling="2025-11-24 18:20:45.493033166 +0000 UTC m=+3238.090700968" lastFinishedPulling="2025-11-24 18:20:46.948835386 +0000 UTC m=+3239.546503218" observedRunningTime="2025-11-24 18:20:47.80328137 +0000 UTC m=+3240.400949212" watchObservedRunningTime="2025-11-24 18:20:47.810740039 +0000 UTC m=+3240.408407881" Nov 24 18:20:50 crc kubenswrapper[4808]: I1124 18:20:50.347208 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:20:50 crc kubenswrapper[4808]: E1124 18:20:50.348164 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:21:05 crc kubenswrapper[4808]: I1124 18:21:05.348600 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:21:05 crc kubenswrapper[4808]: E1124 18:21:05.349605 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:21:09 crc kubenswrapper[4808]: I1124 18:21:09.757946 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bxz78/must-gather-tbb5f"] Nov 24 18:21:09 crc kubenswrapper[4808]: I1124 18:21:09.765060 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/must-gather-tbb5f" Nov 24 18:21:09 crc kubenswrapper[4808]: I1124 18:21:09.767965 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bxz78"/"kube-root-ca.crt" Nov 24 18:21:09 crc kubenswrapper[4808]: I1124 18:21:09.768271 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bxz78"/"openshift-service-ca.crt" Nov 24 18:21:09 crc kubenswrapper[4808]: I1124 18:21:09.788960 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bxz78/must-gather-tbb5f"] Nov 24 18:21:09 crc kubenswrapper[4808]: I1124 18:21:09.901050 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b3d20d4f-5b20-4f82-8e89-f232fd4abfbe-must-gather-output\") pod \"must-gather-tbb5f\" (UID: \"b3d20d4f-5b20-4f82-8e89-f232fd4abfbe\") " pod="openshift-must-gather-bxz78/must-gather-tbb5f" Nov 24 18:21:09 crc kubenswrapper[4808]: I1124 18:21:09.901455 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stfkb\" (UniqueName: \"kubernetes.io/projected/b3d20d4f-5b20-4f82-8e89-f232fd4abfbe-kube-api-access-stfkb\") pod \"must-gather-tbb5f\" (UID: \"b3d20d4f-5b20-4f82-8e89-f232fd4abfbe\") " pod="openshift-must-gather-bxz78/must-gather-tbb5f" Nov 24 18:21:10 crc kubenswrapper[4808]: I1124 18:21:10.003703 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b3d20d4f-5b20-4f82-8e89-f232fd4abfbe-must-gather-output\") pod \"must-gather-tbb5f\" (UID: \"b3d20d4f-5b20-4f82-8e89-f232fd4abfbe\") " pod="openshift-must-gather-bxz78/must-gather-tbb5f" Nov 24 18:21:10 crc kubenswrapper[4808]: I1124 18:21:10.003856 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stfkb\" (UniqueName: \"kubernetes.io/projected/b3d20d4f-5b20-4f82-8e89-f232fd4abfbe-kube-api-access-stfkb\") pod \"must-gather-tbb5f\" (UID: \"b3d20d4f-5b20-4f82-8e89-f232fd4abfbe\") " pod="openshift-must-gather-bxz78/must-gather-tbb5f" Nov 24 18:21:10 crc kubenswrapper[4808]: I1124 18:21:10.004326 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b3d20d4f-5b20-4f82-8e89-f232fd4abfbe-must-gather-output\") pod \"must-gather-tbb5f\" (UID: \"b3d20d4f-5b20-4f82-8e89-f232fd4abfbe\") " pod="openshift-must-gather-bxz78/must-gather-tbb5f" Nov 24 18:21:10 crc kubenswrapper[4808]: I1124 18:21:10.026970 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stfkb\" (UniqueName: \"kubernetes.io/projected/b3d20d4f-5b20-4f82-8e89-f232fd4abfbe-kube-api-access-stfkb\") pod \"must-gather-tbb5f\" (UID: \"b3d20d4f-5b20-4f82-8e89-f232fd4abfbe\") " pod="openshift-must-gather-bxz78/must-gather-tbb5f" Nov 24 18:21:10 crc kubenswrapper[4808]: I1124 18:21:10.113640 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/must-gather-tbb5f" Nov 24 18:21:10 crc kubenswrapper[4808]: I1124 18:21:10.613111 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bxz78/must-gather-tbb5f"] Nov 24 18:21:10 crc kubenswrapper[4808]: W1124 18:21:10.619637 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3d20d4f_5b20_4f82_8e89_f232fd4abfbe.slice/crio-dd555956f12f0df11312b3713a874af659f125bd10943888df41f6e1ad115cff WatchSource:0}: Error finding container dd555956f12f0df11312b3713a874af659f125bd10943888df41f6e1ad115cff: Status 404 returned error can't find the container with id dd555956f12f0df11312b3713a874af659f125bd10943888df41f6e1ad115cff Nov 24 18:21:11 crc kubenswrapper[4808]: I1124 18:21:11.062128 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bxz78/must-gather-tbb5f" event={"ID":"b3d20d4f-5b20-4f82-8e89-f232fd4abfbe","Type":"ContainerStarted","Data":"dd555956f12f0df11312b3713a874af659f125bd10943888df41f6e1ad115cff"} Nov 24 18:21:15 crc kubenswrapper[4808]: I1124 18:21:15.101633 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bxz78/must-gather-tbb5f" event={"ID":"b3d20d4f-5b20-4f82-8e89-f232fd4abfbe","Type":"ContainerStarted","Data":"e71fec119fe999e3b12414f1e1f0b907fde36af74c69b188eab679872232a824"} Nov 24 18:21:16 crc kubenswrapper[4808]: I1124 18:21:16.117470 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bxz78/must-gather-tbb5f" event={"ID":"b3d20d4f-5b20-4f82-8e89-f232fd4abfbe","Type":"ContainerStarted","Data":"70ecfba37d6166ed8dc67a6787df0f1d522a6b092bd4902f4fb542fc1bd46c09"} Nov 24 18:21:16 crc kubenswrapper[4808]: I1124 18:21:16.142188 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bxz78/must-gather-tbb5f" podStartSLOduration=3.018267459 podStartE2EDuration="7.142160457s" podCreationTimestamp="2025-11-24 18:21:09 +0000 UTC" firstStartedPulling="2025-11-24 18:21:10.621437174 +0000 UTC m=+3263.219104976" lastFinishedPulling="2025-11-24 18:21:14.745330172 +0000 UTC m=+3267.342997974" observedRunningTime="2025-11-24 18:21:16.141411856 +0000 UTC m=+3268.739079688" watchObservedRunningTime="2025-11-24 18:21:16.142160457 +0000 UTC m=+3268.739828299" Nov 24 18:21:18 crc kubenswrapper[4808]: I1124 18:21:18.683090 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bxz78/crc-debug-68tnz"] Nov 24 18:21:18 crc kubenswrapper[4808]: I1124 18:21:18.684823 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/crc-debug-68tnz" Nov 24 18:21:18 crc kubenswrapper[4808]: I1124 18:21:18.687237 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-bxz78"/"default-dockercfg-2h5mn" Nov 24 18:21:18 crc kubenswrapper[4808]: I1124 18:21:18.780938 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brt8v\" (UniqueName: \"kubernetes.io/projected/fdf61447-fa63-4c6c-a4d7-ef78f9c7854c-kube-api-access-brt8v\") pod \"crc-debug-68tnz\" (UID: \"fdf61447-fa63-4c6c-a4d7-ef78f9c7854c\") " pod="openshift-must-gather-bxz78/crc-debug-68tnz" Nov 24 18:21:18 crc kubenswrapper[4808]: I1124 18:21:18.781003 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fdf61447-fa63-4c6c-a4d7-ef78f9c7854c-host\") pod \"crc-debug-68tnz\" (UID: \"fdf61447-fa63-4c6c-a4d7-ef78f9c7854c\") " pod="openshift-must-gather-bxz78/crc-debug-68tnz" Nov 24 18:21:18 crc kubenswrapper[4808]: I1124 18:21:18.883353 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brt8v\" (UniqueName: \"kubernetes.io/projected/fdf61447-fa63-4c6c-a4d7-ef78f9c7854c-kube-api-access-brt8v\") pod \"crc-debug-68tnz\" (UID: \"fdf61447-fa63-4c6c-a4d7-ef78f9c7854c\") " pod="openshift-must-gather-bxz78/crc-debug-68tnz" Nov 24 18:21:18 crc kubenswrapper[4808]: I1124 18:21:18.883407 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fdf61447-fa63-4c6c-a4d7-ef78f9c7854c-host\") pod \"crc-debug-68tnz\" (UID: \"fdf61447-fa63-4c6c-a4d7-ef78f9c7854c\") " pod="openshift-must-gather-bxz78/crc-debug-68tnz" Nov 24 18:21:18 crc kubenswrapper[4808]: I1124 18:21:18.883635 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fdf61447-fa63-4c6c-a4d7-ef78f9c7854c-host\") pod \"crc-debug-68tnz\" (UID: \"fdf61447-fa63-4c6c-a4d7-ef78f9c7854c\") " pod="openshift-must-gather-bxz78/crc-debug-68tnz" Nov 24 18:21:18 crc kubenswrapper[4808]: I1124 18:21:18.916686 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brt8v\" (UniqueName: \"kubernetes.io/projected/fdf61447-fa63-4c6c-a4d7-ef78f9c7854c-kube-api-access-brt8v\") pod \"crc-debug-68tnz\" (UID: \"fdf61447-fa63-4c6c-a4d7-ef78f9c7854c\") " pod="openshift-must-gather-bxz78/crc-debug-68tnz" Nov 24 18:21:19 crc kubenswrapper[4808]: I1124 18:21:19.002660 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/crc-debug-68tnz" Nov 24 18:21:19 crc kubenswrapper[4808]: I1124 18:21:19.146850 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bxz78/crc-debug-68tnz" event={"ID":"fdf61447-fa63-4c6c-a4d7-ef78f9c7854c","Type":"ContainerStarted","Data":"364716a9330223f132b3310ec5033fe432b96d67cf0bd99e90ac946caa06ead6"} Nov 24 18:21:20 crc kubenswrapper[4808]: I1124 18:21:20.351196 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:21:20 crc kubenswrapper[4808]: E1124 18:21:20.351761 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:21:30 crc kubenswrapper[4808]: I1124 18:21:30.245582 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bxz78/crc-debug-68tnz" event={"ID":"fdf61447-fa63-4c6c-a4d7-ef78f9c7854c","Type":"ContainerStarted","Data":"2f7ef1f21801221f023f341824b577cbe61fb18a26a475ad4a1cbedcfe60e0f5"} Nov 24 18:21:30 crc kubenswrapper[4808]: I1124 18:21:30.272585 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bxz78/crc-debug-68tnz" podStartSLOduration=2.204652634 podStartE2EDuration="12.272559765s" podCreationTimestamp="2025-11-24 18:21:18 +0000 UTC" firstStartedPulling="2025-11-24 18:21:19.0703323 +0000 UTC m=+3271.668000102" lastFinishedPulling="2025-11-24 18:21:29.138239411 +0000 UTC m=+3281.735907233" observedRunningTime="2025-11-24 18:21:30.260114216 +0000 UTC m=+3282.857782028" watchObservedRunningTime="2025-11-24 18:21:30.272559765 +0000 UTC m=+3282.870227587" Nov 24 18:21:34 crc kubenswrapper[4808]: I1124 18:21:34.347816 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:21:34 crc kubenswrapper[4808]: E1124 18:21:34.348542 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:21:46 crc kubenswrapper[4808]: I1124 18:21:46.146096 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wzlml"] Nov 24 18:21:46 crc kubenswrapper[4808]: I1124 18:21:46.151909 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:21:46 crc kubenswrapper[4808]: I1124 18:21:46.163887 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wzlml"] Nov 24 18:21:46 crc kubenswrapper[4808]: I1124 18:21:46.303102 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfqz7\" (UniqueName: \"kubernetes.io/projected/437cdcc2-0577-4a24-a9a5-87d5394cba03-kube-api-access-lfqz7\") pod \"redhat-operators-wzlml\" (UID: \"437cdcc2-0577-4a24-a9a5-87d5394cba03\") " pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:21:46 crc kubenswrapper[4808]: I1124 18:21:46.303189 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/437cdcc2-0577-4a24-a9a5-87d5394cba03-utilities\") pod \"redhat-operators-wzlml\" (UID: \"437cdcc2-0577-4a24-a9a5-87d5394cba03\") " pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:21:46 crc kubenswrapper[4808]: I1124 18:21:46.303247 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/437cdcc2-0577-4a24-a9a5-87d5394cba03-catalog-content\") pod \"redhat-operators-wzlml\" (UID: \"437cdcc2-0577-4a24-a9a5-87d5394cba03\") " pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:21:46 crc kubenswrapper[4808]: I1124 18:21:46.347336 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:21:46 crc kubenswrapper[4808]: E1124 18:21:46.347607 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:21:46 crc kubenswrapper[4808]: I1124 18:21:46.404821 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/437cdcc2-0577-4a24-a9a5-87d5394cba03-utilities\") pod \"redhat-operators-wzlml\" (UID: \"437cdcc2-0577-4a24-a9a5-87d5394cba03\") " pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:21:46 crc kubenswrapper[4808]: I1124 18:21:46.404935 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/437cdcc2-0577-4a24-a9a5-87d5394cba03-catalog-content\") pod \"redhat-operators-wzlml\" (UID: \"437cdcc2-0577-4a24-a9a5-87d5394cba03\") " pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:21:46 crc kubenswrapper[4808]: I1124 18:21:46.405082 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfqz7\" (UniqueName: \"kubernetes.io/projected/437cdcc2-0577-4a24-a9a5-87d5394cba03-kube-api-access-lfqz7\") pod \"redhat-operators-wzlml\" (UID: \"437cdcc2-0577-4a24-a9a5-87d5394cba03\") " pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:21:46 crc kubenswrapper[4808]: I1124 18:21:46.405414 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/437cdcc2-0577-4a24-a9a5-87d5394cba03-utilities\") pod \"redhat-operators-wzlml\" (UID: \"437cdcc2-0577-4a24-a9a5-87d5394cba03\") " pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:21:46 crc kubenswrapper[4808]: I1124 18:21:46.405448 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/437cdcc2-0577-4a24-a9a5-87d5394cba03-catalog-content\") pod \"redhat-operators-wzlml\" (UID: \"437cdcc2-0577-4a24-a9a5-87d5394cba03\") " pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:21:46 crc kubenswrapper[4808]: I1124 18:21:46.431770 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfqz7\" (UniqueName: \"kubernetes.io/projected/437cdcc2-0577-4a24-a9a5-87d5394cba03-kube-api-access-lfqz7\") pod \"redhat-operators-wzlml\" (UID: \"437cdcc2-0577-4a24-a9a5-87d5394cba03\") " pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:21:46 crc kubenswrapper[4808]: I1124 18:21:46.480284 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:21:46 crc kubenswrapper[4808]: I1124 18:21:46.992198 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wzlml"] Nov 24 18:21:47 crc kubenswrapper[4808]: W1124 18:21:47.039091 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod437cdcc2_0577_4a24_a9a5_87d5394cba03.slice/crio-f48e44a4de502836e68c2ffc1b182aae2203d861f74d7f118604b8af7046e985 WatchSource:0}: Error finding container f48e44a4de502836e68c2ffc1b182aae2203d861f74d7f118604b8af7046e985: Status 404 returned error can't find the container with id f48e44a4de502836e68c2ffc1b182aae2203d861f74d7f118604b8af7046e985 Nov 24 18:21:47 crc kubenswrapper[4808]: I1124 18:21:47.403195 4808 generic.go:334] "Generic (PLEG): container finished" podID="437cdcc2-0577-4a24-a9a5-87d5394cba03" containerID="e958927137424e2121ca4f494ce4a61a34865d9bb28675bf5b3318642bd4f04f" exitCode=0 Nov 24 18:21:47 crc kubenswrapper[4808]: I1124 18:21:47.403279 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzlml" event={"ID":"437cdcc2-0577-4a24-a9a5-87d5394cba03","Type":"ContainerDied","Data":"e958927137424e2121ca4f494ce4a61a34865d9bb28675bf5b3318642bd4f04f"} Nov 24 18:21:47 crc kubenswrapper[4808]: I1124 18:21:47.403516 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzlml" event={"ID":"437cdcc2-0577-4a24-a9a5-87d5394cba03","Type":"ContainerStarted","Data":"f48e44a4de502836e68c2ffc1b182aae2203d861f74d7f118604b8af7046e985"} Nov 24 18:21:49 crc kubenswrapper[4808]: I1124 18:21:49.424053 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzlml" event={"ID":"437cdcc2-0577-4a24-a9a5-87d5394cba03","Type":"ContainerStarted","Data":"f961474e4e12fd3864abae681a47188f9ff523f1e0b5d4a41ebf4b7cc682a6f6"} Nov 24 18:21:50 crc kubenswrapper[4808]: I1124 18:21:50.432793 4808 generic.go:334] "Generic (PLEG): container finished" podID="437cdcc2-0577-4a24-a9a5-87d5394cba03" containerID="f961474e4e12fd3864abae681a47188f9ff523f1e0b5d4a41ebf4b7cc682a6f6" exitCode=0 Nov 24 18:21:50 crc kubenswrapper[4808]: I1124 18:21:50.432841 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzlml" event={"ID":"437cdcc2-0577-4a24-a9a5-87d5394cba03","Type":"ContainerDied","Data":"f961474e4e12fd3864abae681a47188f9ff523f1e0b5d4a41ebf4b7cc682a6f6"} Nov 24 18:21:51 crc kubenswrapper[4808]: I1124 18:21:51.443647 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzlml" event={"ID":"437cdcc2-0577-4a24-a9a5-87d5394cba03","Type":"ContainerStarted","Data":"cba65946cfb27bddfc1f0d550813f00b605be41c9e20be012799dc773684048a"} Nov 24 18:21:51 crc kubenswrapper[4808]: I1124 18:21:51.461972 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wzlml" podStartSLOduration=2.057174409 podStartE2EDuration="5.461951518s" podCreationTimestamp="2025-11-24 18:21:46 +0000 UTC" firstStartedPulling="2025-11-24 18:21:47.426516401 +0000 UTC m=+3300.024184203" lastFinishedPulling="2025-11-24 18:21:50.83129351 +0000 UTC m=+3303.428961312" observedRunningTime="2025-11-24 18:21:51.460779705 +0000 UTC m=+3304.058447507" watchObservedRunningTime="2025-11-24 18:21:51.461951518 +0000 UTC m=+3304.059619320" Nov 24 18:21:56 crc kubenswrapper[4808]: I1124 18:21:56.480735 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:21:56 crc kubenswrapper[4808]: I1124 18:21:56.481529 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:21:57 crc kubenswrapper[4808]: I1124 18:21:57.558075 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wzlml" podUID="437cdcc2-0577-4a24-a9a5-87d5394cba03" containerName="registry-server" probeResult="failure" output=< Nov 24 18:21:57 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Nov 24 18:21:57 crc kubenswrapper[4808]: > Nov 24 18:22:01 crc kubenswrapper[4808]: I1124 18:22:01.347867 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:22:01 crc kubenswrapper[4808]: E1124 18:22:01.348711 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:22:06 crc kubenswrapper[4808]: I1124 18:22:06.560918 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:22:06 crc kubenswrapper[4808]: I1124 18:22:06.623632 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:22:06 crc kubenswrapper[4808]: I1124 18:22:06.810280 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wzlml"] Nov 24 18:22:07 crc kubenswrapper[4808]: I1124 18:22:07.608722 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wzlml" podUID="437cdcc2-0577-4a24-a9a5-87d5394cba03" containerName="registry-server" containerID="cri-o://cba65946cfb27bddfc1f0d550813f00b605be41c9e20be012799dc773684048a" gracePeriod=2 Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.124994 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.157438 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/437cdcc2-0577-4a24-a9a5-87d5394cba03-utilities\") pod \"437cdcc2-0577-4a24-a9a5-87d5394cba03\" (UID: \"437cdcc2-0577-4a24-a9a5-87d5394cba03\") " Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.157489 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfqz7\" (UniqueName: \"kubernetes.io/projected/437cdcc2-0577-4a24-a9a5-87d5394cba03-kube-api-access-lfqz7\") pod \"437cdcc2-0577-4a24-a9a5-87d5394cba03\" (UID: \"437cdcc2-0577-4a24-a9a5-87d5394cba03\") " Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.157576 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/437cdcc2-0577-4a24-a9a5-87d5394cba03-catalog-content\") pod \"437cdcc2-0577-4a24-a9a5-87d5394cba03\" (UID: \"437cdcc2-0577-4a24-a9a5-87d5394cba03\") " Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.159270 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/437cdcc2-0577-4a24-a9a5-87d5394cba03-utilities" (OuterVolumeSpecName: "utilities") pod "437cdcc2-0577-4a24-a9a5-87d5394cba03" (UID: "437cdcc2-0577-4a24-a9a5-87d5394cba03"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.188584 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/437cdcc2-0577-4a24-a9a5-87d5394cba03-kube-api-access-lfqz7" (OuterVolumeSpecName: "kube-api-access-lfqz7") pod "437cdcc2-0577-4a24-a9a5-87d5394cba03" (UID: "437cdcc2-0577-4a24-a9a5-87d5394cba03"). InnerVolumeSpecName "kube-api-access-lfqz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.260928 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/437cdcc2-0577-4a24-a9a5-87d5394cba03-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.260979 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfqz7\" (UniqueName: \"kubernetes.io/projected/437cdcc2-0577-4a24-a9a5-87d5394cba03-kube-api-access-lfqz7\") on node \"crc\" DevicePath \"\"" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.293888 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/437cdcc2-0577-4a24-a9a5-87d5394cba03-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "437cdcc2-0577-4a24-a9a5-87d5394cba03" (UID: "437cdcc2-0577-4a24-a9a5-87d5394cba03"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.362831 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/437cdcc2-0577-4a24-a9a5-87d5394cba03-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.620635 4808 generic.go:334] "Generic (PLEG): container finished" podID="437cdcc2-0577-4a24-a9a5-87d5394cba03" containerID="cba65946cfb27bddfc1f0d550813f00b605be41c9e20be012799dc773684048a" exitCode=0 Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.620688 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzlml" event={"ID":"437cdcc2-0577-4a24-a9a5-87d5394cba03","Type":"ContainerDied","Data":"cba65946cfb27bddfc1f0d550813f00b605be41c9e20be012799dc773684048a"} Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.620720 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzlml" event={"ID":"437cdcc2-0577-4a24-a9a5-87d5394cba03","Type":"ContainerDied","Data":"f48e44a4de502836e68c2ffc1b182aae2203d861f74d7f118604b8af7046e985"} Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.620741 4808 scope.go:117] "RemoveContainer" containerID="cba65946cfb27bddfc1f0d550813f00b605be41c9e20be012799dc773684048a" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.620740 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wzlml" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.658108 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wzlml"] Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.663859 4808 scope.go:117] "RemoveContainer" containerID="f961474e4e12fd3864abae681a47188f9ff523f1e0b5d4a41ebf4b7cc682a6f6" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.669285 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wzlml"] Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.698715 4808 scope.go:117] "RemoveContainer" containerID="e958927137424e2121ca4f494ce4a61a34865d9bb28675bf5b3318642bd4f04f" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.772538 4808 scope.go:117] "RemoveContainer" containerID="cba65946cfb27bddfc1f0d550813f00b605be41c9e20be012799dc773684048a" Nov 24 18:22:08 crc kubenswrapper[4808]: E1124 18:22:08.773004 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cba65946cfb27bddfc1f0d550813f00b605be41c9e20be012799dc773684048a\": container with ID starting with cba65946cfb27bddfc1f0d550813f00b605be41c9e20be012799dc773684048a not found: ID does not exist" containerID="cba65946cfb27bddfc1f0d550813f00b605be41c9e20be012799dc773684048a" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.773062 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cba65946cfb27bddfc1f0d550813f00b605be41c9e20be012799dc773684048a"} err="failed to get container status \"cba65946cfb27bddfc1f0d550813f00b605be41c9e20be012799dc773684048a\": rpc error: code = NotFound desc = could not find container \"cba65946cfb27bddfc1f0d550813f00b605be41c9e20be012799dc773684048a\": container with ID starting with cba65946cfb27bddfc1f0d550813f00b605be41c9e20be012799dc773684048a not found: ID does not exist" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.773093 4808 scope.go:117] "RemoveContainer" containerID="f961474e4e12fd3864abae681a47188f9ff523f1e0b5d4a41ebf4b7cc682a6f6" Nov 24 18:22:08 crc kubenswrapper[4808]: E1124 18:22:08.773568 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f961474e4e12fd3864abae681a47188f9ff523f1e0b5d4a41ebf4b7cc682a6f6\": container with ID starting with f961474e4e12fd3864abae681a47188f9ff523f1e0b5d4a41ebf4b7cc682a6f6 not found: ID does not exist" containerID="f961474e4e12fd3864abae681a47188f9ff523f1e0b5d4a41ebf4b7cc682a6f6" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.773605 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f961474e4e12fd3864abae681a47188f9ff523f1e0b5d4a41ebf4b7cc682a6f6"} err="failed to get container status \"f961474e4e12fd3864abae681a47188f9ff523f1e0b5d4a41ebf4b7cc682a6f6\": rpc error: code = NotFound desc = could not find container \"f961474e4e12fd3864abae681a47188f9ff523f1e0b5d4a41ebf4b7cc682a6f6\": container with ID starting with f961474e4e12fd3864abae681a47188f9ff523f1e0b5d4a41ebf4b7cc682a6f6 not found: ID does not exist" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.773618 4808 scope.go:117] "RemoveContainer" containerID="e958927137424e2121ca4f494ce4a61a34865d9bb28675bf5b3318642bd4f04f" Nov 24 18:22:08 crc kubenswrapper[4808]: E1124 18:22:08.773881 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e958927137424e2121ca4f494ce4a61a34865d9bb28675bf5b3318642bd4f04f\": container with ID starting with e958927137424e2121ca4f494ce4a61a34865d9bb28675bf5b3318642bd4f04f not found: ID does not exist" containerID="e958927137424e2121ca4f494ce4a61a34865d9bb28675bf5b3318642bd4f04f" Nov 24 18:22:08 crc kubenswrapper[4808]: I1124 18:22:08.773905 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e958927137424e2121ca4f494ce4a61a34865d9bb28675bf5b3318642bd4f04f"} err="failed to get container status \"e958927137424e2121ca4f494ce4a61a34865d9bb28675bf5b3318642bd4f04f\": rpc error: code = NotFound desc = could not find container \"e958927137424e2121ca4f494ce4a61a34865d9bb28675bf5b3318642bd4f04f\": container with ID starting with e958927137424e2121ca4f494ce4a61a34865d9bb28675bf5b3318642bd4f04f not found: ID does not exist" Nov 24 18:22:09 crc kubenswrapper[4808]: I1124 18:22:09.638707 4808 generic.go:334] "Generic (PLEG): container finished" podID="fdf61447-fa63-4c6c-a4d7-ef78f9c7854c" containerID="2f7ef1f21801221f023f341824b577cbe61fb18a26a475ad4a1cbedcfe60e0f5" exitCode=0 Nov 24 18:22:09 crc kubenswrapper[4808]: I1124 18:22:09.638773 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bxz78/crc-debug-68tnz" event={"ID":"fdf61447-fa63-4c6c-a4d7-ef78f9c7854c","Type":"ContainerDied","Data":"2f7ef1f21801221f023f341824b577cbe61fb18a26a475ad4a1cbedcfe60e0f5"} Nov 24 18:22:10 crc kubenswrapper[4808]: I1124 18:22:10.366726 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="437cdcc2-0577-4a24-a9a5-87d5394cba03" path="/var/lib/kubelet/pods/437cdcc2-0577-4a24-a9a5-87d5394cba03/volumes" Nov 24 18:22:10 crc kubenswrapper[4808]: I1124 18:22:10.786200 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/crc-debug-68tnz" Nov 24 18:22:10 crc kubenswrapper[4808]: I1124 18:22:10.811886 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fdf61447-fa63-4c6c-a4d7-ef78f9c7854c-host\") pod \"fdf61447-fa63-4c6c-a4d7-ef78f9c7854c\" (UID: \"fdf61447-fa63-4c6c-a4d7-ef78f9c7854c\") " Nov 24 18:22:10 crc kubenswrapper[4808]: I1124 18:22:10.812014 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brt8v\" (UniqueName: \"kubernetes.io/projected/fdf61447-fa63-4c6c-a4d7-ef78f9c7854c-kube-api-access-brt8v\") pod \"fdf61447-fa63-4c6c-a4d7-ef78f9c7854c\" (UID: \"fdf61447-fa63-4c6c-a4d7-ef78f9c7854c\") " Nov 24 18:22:10 crc kubenswrapper[4808]: I1124 18:22:10.812067 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fdf61447-fa63-4c6c-a4d7-ef78f9c7854c-host" (OuterVolumeSpecName: "host") pod "fdf61447-fa63-4c6c-a4d7-ef78f9c7854c" (UID: "fdf61447-fa63-4c6c-a4d7-ef78f9c7854c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 18:22:10 crc kubenswrapper[4808]: I1124 18:22:10.812381 4808 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fdf61447-fa63-4c6c-a4d7-ef78f9c7854c-host\") on node \"crc\" DevicePath \"\"" Nov 24 18:22:10 crc kubenswrapper[4808]: I1124 18:22:10.822045 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdf61447-fa63-4c6c-a4d7-ef78f9c7854c-kube-api-access-brt8v" (OuterVolumeSpecName: "kube-api-access-brt8v") pod "fdf61447-fa63-4c6c-a4d7-ef78f9c7854c" (UID: "fdf61447-fa63-4c6c-a4d7-ef78f9c7854c"). InnerVolumeSpecName "kube-api-access-brt8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:22:10 crc kubenswrapper[4808]: I1124 18:22:10.832427 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bxz78/crc-debug-68tnz"] Nov 24 18:22:10 crc kubenswrapper[4808]: I1124 18:22:10.843625 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bxz78/crc-debug-68tnz"] Nov 24 18:22:10 crc kubenswrapper[4808]: I1124 18:22:10.913876 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brt8v\" (UniqueName: \"kubernetes.io/projected/fdf61447-fa63-4c6c-a4d7-ef78f9c7854c-kube-api-access-brt8v\") on node \"crc\" DevicePath \"\"" Nov 24 18:22:11 crc kubenswrapper[4808]: I1124 18:22:11.668103 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="364716a9330223f132b3310ec5033fe432b96d67cf0bd99e90ac946caa06ead6" Nov 24 18:22:11 crc kubenswrapper[4808]: I1124 18:22:11.668303 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/crc-debug-68tnz" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.121263 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bxz78/crc-debug-9v6gp"] Nov 24 18:22:12 crc kubenswrapper[4808]: E1124 18:22:12.121737 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="437cdcc2-0577-4a24-a9a5-87d5394cba03" containerName="extract-utilities" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.121759 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="437cdcc2-0577-4a24-a9a5-87d5394cba03" containerName="extract-utilities" Nov 24 18:22:12 crc kubenswrapper[4808]: E1124 18:22:12.121782 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="437cdcc2-0577-4a24-a9a5-87d5394cba03" containerName="extract-content" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.121790 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="437cdcc2-0577-4a24-a9a5-87d5394cba03" containerName="extract-content" Nov 24 18:22:12 crc kubenswrapper[4808]: E1124 18:22:12.121810 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="437cdcc2-0577-4a24-a9a5-87d5394cba03" containerName="registry-server" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.121817 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="437cdcc2-0577-4a24-a9a5-87d5394cba03" containerName="registry-server" Nov 24 18:22:12 crc kubenswrapper[4808]: E1124 18:22:12.121829 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdf61447-fa63-4c6c-a4d7-ef78f9c7854c" containerName="container-00" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.121836 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdf61447-fa63-4c6c-a4d7-ef78f9c7854c" containerName="container-00" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.122088 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdf61447-fa63-4c6c-a4d7-ef78f9c7854c" containerName="container-00" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.122122 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="437cdcc2-0577-4a24-a9a5-87d5394cba03" containerName="registry-server" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.122796 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/crc-debug-9v6gp" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.125053 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-bxz78"/"default-dockercfg-2h5mn" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.272355 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fsj8\" (UniqueName: \"kubernetes.io/projected/f418bed1-b7ad-4e27-971c-e53a298143e7-kube-api-access-8fsj8\") pod \"crc-debug-9v6gp\" (UID: \"f418bed1-b7ad-4e27-971c-e53a298143e7\") " pod="openshift-must-gather-bxz78/crc-debug-9v6gp" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.272574 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f418bed1-b7ad-4e27-971c-e53a298143e7-host\") pod \"crc-debug-9v6gp\" (UID: \"f418bed1-b7ad-4e27-971c-e53a298143e7\") " pod="openshift-must-gather-bxz78/crc-debug-9v6gp" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.347430 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.360476 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdf61447-fa63-4c6c-a4d7-ef78f9c7854c" path="/var/lib/kubelet/pods/fdf61447-fa63-4c6c-a4d7-ef78f9c7854c/volumes" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.374202 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fsj8\" (UniqueName: \"kubernetes.io/projected/f418bed1-b7ad-4e27-971c-e53a298143e7-kube-api-access-8fsj8\") pod \"crc-debug-9v6gp\" (UID: \"f418bed1-b7ad-4e27-971c-e53a298143e7\") " pod="openshift-must-gather-bxz78/crc-debug-9v6gp" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.374591 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f418bed1-b7ad-4e27-971c-e53a298143e7-host\") pod \"crc-debug-9v6gp\" (UID: \"f418bed1-b7ad-4e27-971c-e53a298143e7\") " pod="openshift-must-gather-bxz78/crc-debug-9v6gp" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.374787 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f418bed1-b7ad-4e27-971c-e53a298143e7-host\") pod \"crc-debug-9v6gp\" (UID: \"f418bed1-b7ad-4e27-971c-e53a298143e7\") " pod="openshift-must-gather-bxz78/crc-debug-9v6gp" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.393620 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fsj8\" (UniqueName: \"kubernetes.io/projected/f418bed1-b7ad-4e27-971c-e53a298143e7-kube-api-access-8fsj8\") pod \"crc-debug-9v6gp\" (UID: \"f418bed1-b7ad-4e27-971c-e53a298143e7\") " pod="openshift-must-gather-bxz78/crc-debug-9v6gp" Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.450551 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/crc-debug-9v6gp" Nov 24 18:22:12 crc kubenswrapper[4808]: W1124 18:22:12.518983 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf418bed1_b7ad_4e27_971c_e53a298143e7.slice/crio-03eabb1b92075170365a115ff6c0c0e64d5774620d6ac9e16e7bb07537809e2c WatchSource:0}: Error finding container 03eabb1b92075170365a115ff6c0c0e64d5774620d6ac9e16e7bb07537809e2c: Status 404 returned error can't find the container with id 03eabb1b92075170365a115ff6c0c0e64d5774620d6ac9e16e7bb07537809e2c Nov 24 18:22:12 crc kubenswrapper[4808]: I1124 18:22:12.683314 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bxz78/crc-debug-9v6gp" event={"ID":"f418bed1-b7ad-4e27-971c-e53a298143e7","Type":"ContainerStarted","Data":"03eabb1b92075170365a115ff6c0c0e64d5774620d6ac9e16e7bb07537809e2c"} Nov 24 18:22:13 crc kubenswrapper[4808]: I1124 18:22:13.703396 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"402d672713f9ca792d3e28c3866699dcd97d837b30faeba8a4bf79ceeb5f226e"} Nov 24 18:22:13 crc kubenswrapper[4808]: I1124 18:22:13.706853 4808 generic.go:334] "Generic (PLEG): container finished" podID="f418bed1-b7ad-4e27-971c-e53a298143e7" containerID="c5e12c68829128c98d4ecda5525cfc5055f460272679478338ec34ba78eaedad" exitCode=0 Nov 24 18:22:13 crc kubenswrapper[4808]: I1124 18:22:13.706908 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bxz78/crc-debug-9v6gp" event={"ID":"f418bed1-b7ad-4e27-971c-e53a298143e7","Type":"ContainerDied","Data":"c5e12c68829128c98d4ecda5525cfc5055f460272679478338ec34ba78eaedad"} Nov 24 18:22:14 crc kubenswrapper[4808]: I1124 18:22:14.269955 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bxz78/crc-debug-9v6gp"] Nov 24 18:22:14 crc kubenswrapper[4808]: I1124 18:22:14.278216 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bxz78/crc-debug-9v6gp"] Nov 24 18:22:14 crc kubenswrapper[4808]: I1124 18:22:14.859718 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/crc-debug-9v6gp" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.033601 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f418bed1-b7ad-4e27-971c-e53a298143e7-host\") pod \"f418bed1-b7ad-4e27-971c-e53a298143e7\" (UID: \"f418bed1-b7ad-4e27-971c-e53a298143e7\") " Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.033737 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fsj8\" (UniqueName: \"kubernetes.io/projected/f418bed1-b7ad-4e27-971c-e53a298143e7-kube-api-access-8fsj8\") pod \"f418bed1-b7ad-4e27-971c-e53a298143e7\" (UID: \"f418bed1-b7ad-4e27-971c-e53a298143e7\") " Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.033730 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f418bed1-b7ad-4e27-971c-e53a298143e7-host" (OuterVolumeSpecName: "host") pod "f418bed1-b7ad-4e27-971c-e53a298143e7" (UID: "f418bed1-b7ad-4e27-971c-e53a298143e7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.034617 4808 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f418bed1-b7ad-4e27-971c-e53a298143e7-host\") on node \"crc\" DevicePath \"\"" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.041056 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f418bed1-b7ad-4e27-971c-e53a298143e7-kube-api-access-8fsj8" (OuterVolumeSpecName: "kube-api-access-8fsj8") pod "f418bed1-b7ad-4e27-971c-e53a298143e7" (UID: "f418bed1-b7ad-4e27-971c-e53a298143e7"). InnerVolumeSpecName "kube-api-access-8fsj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.135716 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fsj8\" (UniqueName: \"kubernetes.io/projected/f418bed1-b7ad-4e27-971c-e53a298143e7-kube-api-access-8fsj8\") on node \"crc\" DevicePath \"\"" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.512304 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bxz78/crc-debug-zm4sl"] Nov 24 18:22:15 crc kubenswrapper[4808]: E1124 18:22:15.512726 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f418bed1-b7ad-4e27-971c-e53a298143e7" containerName="container-00" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.512745 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f418bed1-b7ad-4e27-971c-e53a298143e7" containerName="container-00" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.512998 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f418bed1-b7ad-4e27-971c-e53a298143e7" containerName="container-00" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.514005 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/crc-debug-zm4sl" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.652740 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86rrh\" (UniqueName: \"kubernetes.io/projected/439b11c8-b32f-4e2e-af0b-1889ad66e9fd-kube-api-access-86rrh\") pod \"crc-debug-zm4sl\" (UID: \"439b11c8-b32f-4e2e-af0b-1889ad66e9fd\") " pod="openshift-must-gather-bxz78/crc-debug-zm4sl" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.652917 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/439b11c8-b32f-4e2e-af0b-1889ad66e9fd-host\") pod \"crc-debug-zm4sl\" (UID: \"439b11c8-b32f-4e2e-af0b-1889ad66e9fd\") " pod="openshift-must-gather-bxz78/crc-debug-zm4sl" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.741295 4808 scope.go:117] "RemoveContainer" containerID="c5e12c68829128c98d4ecda5525cfc5055f460272679478338ec34ba78eaedad" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.741351 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/crc-debug-9v6gp" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.755531 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86rrh\" (UniqueName: \"kubernetes.io/projected/439b11c8-b32f-4e2e-af0b-1889ad66e9fd-kube-api-access-86rrh\") pod \"crc-debug-zm4sl\" (UID: \"439b11c8-b32f-4e2e-af0b-1889ad66e9fd\") " pod="openshift-must-gather-bxz78/crc-debug-zm4sl" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.755672 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/439b11c8-b32f-4e2e-af0b-1889ad66e9fd-host\") pod \"crc-debug-zm4sl\" (UID: \"439b11c8-b32f-4e2e-af0b-1889ad66e9fd\") " pod="openshift-must-gather-bxz78/crc-debug-zm4sl" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.755869 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/439b11c8-b32f-4e2e-af0b-1889ad66e9fd-host\") pod \"crc-debug-zm4sl\" (UID: \"439b11c8-b32f-4e2e-af0b-1889ad66e9fd\") " pod="openshift-must-gather-bxz78/crc-debug-zm4sl" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.779966 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86rrh\" (UniqueName: \"kubernetes.io/projected/439b11c8-b32f-4e2e-af0b-1889ad66e9fd-kube-api-access-86rrh\") pod \"crc-debug-zm4sl\" (UID: \"439b11c8-b32f-4e2e-af0b-1889ad66e9fd\") " pod="openshift-must-gather-bxz78/crc-debug-zm4sl" Nov 24 18:22:15 crc kubenswrapper[4808]: I1124 18:22:15.830494 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/crc-debug-zm4sl" Nov 24 18:22:15 crc kubenswrapper[4808]: W1124 18:22:15.885570 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod439b11c8_b32f_4e2e_af0b_1889ad66e9fd.slice/crio-562270e349995620d5f59571b9012392eef4bb3dedc6c2ee0ee78a269a381aaf WatchSource:0}: Error finding container 562270e349995620d5f59571b9012392eef4bb3dedc6c2ee0ee78a269a381aaf: Status 404 returned error can't find the container with id 562270e349995620d5f59571b9012392eef4bb3dedc6c2ee0ee78a269a381aaf Nov 24 18:22:16 crc kubenswrapper[4808]: I1124 18:22:16.370120 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f418bed1-b7ad-4e27-971c-e53a298143e7" path="/var/lib/kubelet/pods/f418bed1-b7ad-4e27-971c-e53a298143e7/volumes" Nov 24 18:22:16 crc kubenswrapper[4808]: I1124 18:22:16.757679 4808 generic.go:334] "Generic (PLEG): container finished" podID="439b11c8-b32f-4e2e-af0b-1889ad66e9fd" containerID="01fc16320f1ee155ee32184cdfc54335af5c657b290d5df76860f47480856cc7" exitCode=0 Nov 24 18:22:16 crc kubenswrapper[4808]: I1124 18:22:16.757896 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bxz78/crc-debug-zm4sl" event={"ID":"439b11c8-b32f-4e2e-af0b-1889ad66e9fd","Type":"ContainerDied","Data":"01fc16320f1ee155ee32184cdfc54335af5c657b290d5df76860f47480856cc7"} Nov 24 18:22:16 crc kubenswrapper[4808]: I1124 18:22:16.758143 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bxz78/crc-debug-zm4sl" event={"ID":"439b11c8-b32f-4e2e-af0b-1889ad66e9fd","Type":"ContainerStarted","Data":"562270e349995620d5f59571b9012392eef4bb3dedc6c2ee0ee78a269a381aaf"} Nov 24 18:22:16 crc kubenswrapper[4808]: I1124 18:22:16.825822 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bxz78/crc-debug-zm4sl"] Nov 24 18:22:16 crc kubenswrapper[4808]: I1124 18:22:16.840989 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bxz78/crc-debug-zm4sl"] Nov 24 18:22:17 crc kubenswrapper[4808]: I1124 18:22:17.870167 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/crc-debug-zm4sl" Nov 24 18:22:18 crc kubenswrapper[4808]: I1124 18:22:18.005972 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/439b11c8-b32f-4e2e-af0b-1889ad66e9fd-host\") pod \"439b11c8-b32f-4e2e-af0b-1889ad66e9fd\" (UID: \"439b11c8-b32f-4e2e-af0b-1889ad66e9fd\") " Nov 24 18:22:18 crc kubenswrapper[4808]: I1124 18:22:18.006048 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86rrh\" (UniqueName: \"kubernetes.io/projected/439b11c8-b32f-4e2e-af0b-1889ad66e9fd-kube-api-access-86rrh\") pod \"439b11c8-b32f-4e2e-af0b-1889ad66e9fd\" (UID: \"439b11c8-b32f-4e2e-af0b-1889ad66e9fd\") " Nov 24 18:22:18 crc kubenswrapper[4808]: I1124 18:22:18.006088 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/439b11c8-b32f-4e2e-af0b-1889ad66e9fd-host" (OuterVolumeSpecName: "host") pod "439b11c8-b32f-4e2e-af0b-1889ad66e9fd" (UID: "439b11c8-b32f-4e2e-af0b-1889ad66e9fd"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 18:22:18 crc kubenswrapper[4808]: I1124 18:22:18.006433 4808 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/439b11c8-b32f-4e2e-af0b-1889ad66e9fd-host\") on node \"crc\" DevicePath \"\"" Nov 24 18:22:18 crc kubenswrapper[4808]: I1124 18:22:18.015852 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/439b11c8-b32f-4e2e-af0b-1889ad66e9fd-kube-api-access-86rrh" (OuterVolumeSpecName: "kube-api-access-86rrh") pod "439b11c8-b32f-4e2e-af0b-1889ad66e9fd" (UID: "439b11c8-b32f-4e2e-af0b-1889ad66e9fd"). InnerVolumeSpecName "kube-api-access-86rrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:22:18 crc kubenswrapper[4808]: I1124 18:22:18.108517 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86rrh\" (UniqueName: \"kubernetes.io/projected/439b11c8-b32f-4e2e-af0b-1889ad66e9fd-kube-api-access-86rrh\") on node \"crc\" DevicePath \"\"" Nov 24 18:22:18 crc kubenswrapper[4808]: I1124 18:22:18.360455 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="439b11c8-b32f-4e2e-af0b-1889ad66e9fd" path="/var/lib/kubelet/pods/439b11c8-b32f-4e2e-af0b-1889ad66e9fd/volumes" Nov 24 18:22:18 crc kubenswrapper[4808]: I1124 18:22:18.777529 4808 scope.go:117] "RemoveContainer" containerID="01fc16320f1ee155ee32184cdfc54335af5c657b290d5df76860f47480856cc7" Nov 24 18:22:18 crc kubenswrapper[4808]: I1124 18:22:18.777549 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/crc-debug-zm4sl" Nov 24 18:22:32 crc kubenswrapper[4808]: I1124 18:22:32.013483 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67f98cdffd-vs7fq_45f1a3c5-b0ad-429f-a2ff-5d19a23ca362/barbican-api/0.log" Nov 24 18:22:32 crc kubenswrapper[4808]: I1124 18:22:32.082093 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67f98cdffd-vs7fq_45f1a3c5-b0ad-429f-a2ff-5d19a23ca362/barbican-api-log/0.log" Nov 24 18:22:32 crc kubenswrapper[4808]: I1124 18:22:32.230290 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7549f65886-b8qq5_add8319e-0677-426b-b094-25b79b7a77e4/barbican-keystone-listener/0.log" Nov 24 18:22:32 crc kubenswrapper[4808]: I1124 18:22:32.271973 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7549f65886-b8qq5_add8319e-0677-426b-b094-25b79b7a77e4/barbican-keystone-listener-log/0.log" Nov 24 18:22:32 crc kubenswrapper[4808]: I1124 18:22:32.610905 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5db8598f6c-zs7cb_c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba/barbican-worker/0.log" Nov 24 18:22:32 crc kubenswrapper[4808]: I1124 18:22:32.619995 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5db8598f6c-zs7cb_c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba/barbican-worker-log/0.log" Nov 24 18:22:32 crc kubenswrapper[4808]: I1124 18:22:32.707929 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr_b6ec6116-79f6-4178-9ebc-bc29c07889b3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:32 crc kubenswrapper[4808]: I1124 18:22:32.856202 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a9b7ee10-3d3d-4755-bf68-0d745a8e4c67/ceilometer-central-agent/0.log" Nov 24 18:22:32 crc kubenswrapper[4808]: I1124 18:22:32.913978 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a9b7ee10-3d3d-4755-bf68-0d745a8e4c67/ceilometer-notification-agent/0.log" Nov 24 18:22:32 crc kubenswrapper[4808]: I1124 18:22:32.972110 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a9b7ee10-3d3d-4755-bf68-0d745a8e4c67/proxy-httpd/0.log" Nov 24 18:22:33 crc kubenswrapper[4808]: I1124 18:22:33.064085 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a9b7ee10-3d3d-4755-bf68-0d745a8e4c67/sg-core/0.log" Nov 24 18:22:33 crc kubenswrapper[4808]: I1124 18:22:33.159146 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_323bb785-3e02-469e-b169-22109303439e/cinder-api-log/0.log" Nov 24 18:22:33 crc kubenswrapper[4808]: I1124 18:22:33.219254 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_323bb785-3e02-469e-b169-22109303439e/cinder-api/0.log" Nov 24 18:22:33 crc kubenswrapper[4808]: I1124 18:22:33.343550 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a/cinder-scheduler/0.log" Nov 24 18:22:33 crc kubenswrapper[4808]: I1124 18:22:33.392664 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a/probe/0.log" Nov 24 18:22:33 crc kubenswrapper[4808]: I1124 18:22:33.546724 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wglmk_91323238-4f1e-4dd3-84f2-6fc8bf252642/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:33 crc kubenswrapper[4808]: I1124 18:22:33.598992 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk_faadd5b0-6e65-467e-b510-090ee77362b4/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:33 crc kubenswrapper[4808]: I1124 18:22:33.756324 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-m2sjz_db145eef-fe79-49c1-9fd1-570b506b3367/init/0.log" Nov 24 18:22:33 crc kubenswrapper[4808]: I1124 18:22:33.985837 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-m2sjz_db145eef-fe79-49c1-9fd1-570b506b3367/init/0.log" Nov 24 18:22:33 crc kubenswrapper[4808]: I1124 18:22:33.998454 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-s4f85_ee20a194-4626-4fe3-8138-e103a44a2006/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:34 crc kubenswrapper[4808]: I1124 18:22:34.023207 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-m2sjz_db145eef-fe79-49c1-9fd1-570b506b3367/dnsmasq-dns/0.log" Nov 24 18:22:34 crc kubenswrapper[4808]: I1124 18:22:34.200976 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_dc25c6d7-a302-4ea0-bd25-2b19f9b17a71/glance-httpd/0.log" Nov 24 18:22:34 crc kubenswrapper[4808]: I1124 18:22:34.261251 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_dc25c6d7-a302-4ea0-bd25-2b19f9b17a71/glance-log/0.log" Nov 24 18:22:34 crc kubenswrapper[4808]: I1124 18:22:34.370618 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_88cb8610-eb5e-45de-8028-37fc361beaac/glance-log/0.log" Nov 24 18:22:34 crc kubenswrapper[4808]: I1124 18:22:34.400370 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_88cb8610-eb5e-45de-8028-37fc361beaac/glance-httpd/0.log" Nov 24 18:22:34 crc kubenswrapper[4808]: I1124 18:22:34.586486 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5ddbdb6df8-pv5zj_96142f0a-3231-4617-bbd2-9f440c73908e/horizon/0.log" Nov 24 18:22:34 crc kubenswrapper[4808]: I1124 18:22:34.784407 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf_91e767d6-f4dd-4bf8-b58d-37dc9073e90f/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:34 crc kubenswrapper[4808]: I1124 18:22:34.930360 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5ddbdb6df8-pv5zj_96142f0a-3231-4617-bbd2-9f440c73908e/horizon-log/0.log" Nov 24 18:22:34 crc kubenswrapper[4808]: I1124 18:22:34.992669 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-f8mg9_37c9ab78-83a5-4432-8c14-b1e973891168/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:35 crc kubenswrapper[4808]: I1124 18:22:35.281594 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29400121-9z8sf_f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e/keystone-cron/0.log" Nov 24 18:22:35 crc kubenswrapper[4808]: I1124 18:22:35.284116 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7b7d8889-c8cpc_3c94ac11-4a54-4fde-b2dd-e9e92d58894f/keystone-api/0.log" Nov 24 18:22:35 crc kubenswrapper[4808]: I1124 18:22:35.375344 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_0c8560e5-eccf-482d-bc1b-74e44ccfeff1/kube-state-metrics/0.log" Nov 24 18:22:35 crc kubenswrapper[4808]: I1124 18:22:35.473858 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2_3a26fcdc-f080-4677-a308-b08584153734/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:35 crc kubenswrapper[4808]: I1124 18:22:35.943759 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-ff547b7f9-dfqk9_76a5234c-3951-4c26-92c2-0ead15585dc5/neutron-httpd/0.log" Nov 24 18:22:35 crc kubenswrapper[4808]: I1124 18:22:35.979345 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-ff547b7f9-dfqk9_76a5234c-3951-4c26-92c2-0ead15585dc5/neutron-api/0.log" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.046190 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh_40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.132742 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gbqvb"] Nov 24 18:22:36 crc kubenswrapper[4808]: E1124 18:22:36.134280 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="439b11c8-b32f-4e2e-af0b-1889ad66e9fd" containerName="container-00" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.134300 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="439b11c8-b32f-4e2e-af0b-1889ad66e9fd" containerName="container-00" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.134514 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="439b11c8-b32f-4e2e-af0b-1889ad66e9fd" containerName="container-00" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.136284 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.142128 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gbqvb"] Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.239374 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6bcf\" (UniqueName: \"kubernetes.io/projected/c692a6b8-8b25-4a08-83bd-c0567542b35f-kube-api-access-t6bcf\") pod \"community-operators-gbqvb\" (UID: \"c692a6b8-8b25-4a08-83bd-c0567542b35f\") " pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.239430 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c692a6b8-8b25-4a08-83bd-c0567542b35f-utilities\") pod \"community-operators-gbqvb\" (UID: \"c692a6b8-8b25-4a08-83bd-c0567542b35f\") " pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.239713 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c692a6b8-8b25-4a08-83bd-c0567542b35f-catalog-content\") pod \"community-operators-gbqvb\" (UID: \"c692a6b8-8b25-4a08-83bd-c0567542b35f\") " pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.341454 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c692a6b8-8b25-4a08-83bd-c0567542b35f-catalog-content\") pod \"community-operators-gbqvb\" (UID: \"c692a6b8-8b25-4a08-83bd-c0567542b35f\") " pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.341537 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6bcf\" (UniqueName: \"kubernetes.io/projected/c692a6b8-8b25-4a08-83bd-c0567542b35f-kube-api-access-t6bcf\") pod \"community-operators-gbqvb\" (UID: \"c692a6b8-8b25-4a08-83bd-c0567542b35f\") " pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.341570 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c692a6b8-8b25-4a08-83bd-c0567542b35f-utilities\") pod \"community-operators-gbqvb\" (UID: \"c692a6b8-8b25-4a08-83bd-c0567542b35f\") " pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.342188 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c692a6b8-8b25-4a08-83bd-c0567542b35f-catalog-content\") pod \"community-operators-gbqvb\" (UID: \"c692a6b8-8b25-4a08-83bd-c0567542b35f\") " pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.342221 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c692a6b8-8b25-4a08-83bd-c0567542b35f-utilities\") pod \"community-operators-gbqvb\" (UID: \"c692a6b8-8b25-4a08-83bd-c0567542b35f\") " pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.381461 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6bcf\" (UniqueName: \"kubernetes.io/projected/c692a6b8-8b25-4a08-83bd-c0567542b35f-kube-api-access-t6bcf\") pod \"community-operators-gbqvb\" (UID: \"c692a6b8-8b25-4a08-83bd-c0567542b35f\") " pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.516923 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.526692 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2f6861ff-4148-4476-8a97-ef9228b91c8b/nova-api-log/0.log" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.545753 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_1c4e90b9-4ce7-4931-922f-ce473cc9d5ed/nova-cell0-conductor-conductor/0.log" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.705939 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2f6861ff-4148-4476-8a97-ef9228b91c8b/nova-api-api/0.log" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.962656 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_7074b2db-ae2b-4fee-a890-6cc64cd46cf8/nova-cell1-conductor-conductor/0.log" Nov 24 18:22:36 crc kubenswrapper[4808]: I1124 18:22:36.992610 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gbqvb"] Nov 24 18:22:37 crc kubenswrapper[4808]: I1124 18:22:37.022851 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_d32aec99-9b13-4e2e-af1a-9b55ab86026b/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 18:22:37 crc kubenswrapper[4808]: I1124 18:22:37.327557 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-lfxmh_e03cdbd6-8f86-468f-a798-961c6a4920d7/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:37 crc kubenswrapper[4808]: I1124 18:22:37.355747 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e12cb986-dbb8-42c6-a6f4-7d49d49107d6/nova-metadata-log/0.log" Nov 24 18:22:37 crc kubenswrapper[4808]: I1124 18:22:37.647520 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_73a668e5-4465-455b-a109-3fcee8ea831f/mysql-bootstrap/0.log" Nov 24 18:22:37 crc kubenswrapper[4808]: I1124 18:22:37.657066 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1631f202-c106-4965-874f-7af292288025/nova-scheduler-scheduler/0.log" Nov 24 18:22:37 crc kubenswrapper[4808]: I1124 18:22:37.871897 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_73a668e5-4465-455b-a109-3fcee8ea831f/mysql-bootstrap/0.log" Nov 24 18:22:37 crc kubenswrapper[4808]: I1124 18:22:37.878346 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_73a668e5-4465-455b-a109-3fcee8ea831f/galera/0.log" Nov 24 18:22:37 crc kubenswrapper[4808]: I1124 18:22:37.949780 4808 generic.go:334] "Generic (PLEG): container finished" podID="c692a6b8-8b25-4a08-83bd-c0567542b35f" containerID="a63399574a962f8f3514ef32013b6e9449aeef3c0cd9532093b8f78546e130ba" exitCode=0 Nov 24 18:22:37 crc kubenswrapper[4808]: I1124 18:22:37.949826 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbqvb" event={"ID":"c692a6b8-8b25-4a08-83bd-c0567542b35f","Type":"ContainerDied","Data":"a63399574a962f8f3514ef32013b6e9449aeef3c0cd9532093b8f78546e130ba"} Nov 24 18:22:37 crc kubenswrapper[4808]: I1124 18:22:37.949851 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbqvb" event={"ID":"c692a6b8-8b25-4a08-83bd-c0567542b35f","Type":"ContainerStarted","Data":"1e6150100a2f1bcb780fc2b5e53f9512930e79eefac57ea05e2303a21db7a0a4"} Nov 24 18:22:38 crc kubenswrapper[4808]: I1124 18:22:38.088253 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ea9156df-b46b-43f4-97b8-6f1aaf74b6f6/mysql-bootstrap/0.log" Nov 24 18:22:38 crc kubenswrapper[4808]: I1124 18:22:38.233415 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ea9156df-b46b-43f4-97b8-6f1aaf74b6f6/mysql-bootstrap/0.log" Nov 24 18:22:38 crc kubenswrapper[4808]: I1124 18:22:38.252921 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ea9156df-b46b-43f4-97b8-6f1aaf74b6f6/galera/0.log" Nov 24 18:22:38 crc kubenswrapper[4808]: I1124 18:22:38.292847 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e12cb986-dbb8-42c6-a6f4-7d49d49107d6/nova-metadata-metadata/0.log" Nov 24 18:22:38 crc kubenswrapper[4808]: I1124 18:22:38.471392 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c/openstackclient/0.log" Nov 24 18:22:38 crc kubenswrapper[4808]: I1124 18:22:38.521150 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kp2db_a3ac52cd-4f6a-4360-867b-8d0a156089b4/openstack-network-exporter/0.log" Nov 24 18:22:38 crc kubenswrapper[4808]: I1124 18:22:38.852480 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-58sfc_e091dc40-a06a-482a-873c-7be28cd7fd29/ovsdb-server-init/0.log" Nov 24 18:22:38 crc kubenswrapper[4808]: I1124 18:22:38.972381 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbqvb" event={"ID":"c692a6b8-8b25-4a08-83bd-c0567542b35f","Type":"ContainerStarted","Data":"c1b63456b71f0278173a81de37777d66324e12ceaac195739b1b919775a5ac86"} Nov 24 18:22:39 crc kubenswrapper[4808]: I1124 18:22:39.010901 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-58sfc_e091dc40-a06a-482a-873c-7be28cd7fd29/ovs-vswitchd/0.log" Nov 24 18:22:39 crc kubenswrapper[4808]: I1124 18:22:39.016148 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-58sfc_e091dc40-a06a-482a-873c-7be28cd7fd29/ovsdb-server-init/0.log" Nov 24 18:22:39 crc kubenswrapper[4808]: I1124 18:22:39.097812 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-58sfc_e091dc40-a06a-482a-873c-7be28cd7fd29/ovsdb-server/0.log" Nov 24 18:22:39 crc kubenswrapper[4808]: I1124 18:22:39.224583 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-zrsgl_dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec/ovn-controller/0.log" Nov 24 18:22:39 crc kubenswrapper[4808]: I1124 18:22:39.348782 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-v5zl8_119c4dcd-dc15-44aa-bfde-1a2931d8b83d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:39 crc kubenswrapper[4808]: I1124 18:22:39.439002 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a8ca4d50-8de1-4d3a-b5ed-313d024d224a/openstack-network-exporter/0.log" Nov 24 18:22:39 crc kubenswrapper[4808]: I1124 18:22:39.446405 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a8ca4d50-8de1-4d3a-b5ed-313d024d224a/ovn-northd/0.log" Nov 24 18:22:39 crc kubenswrapper[4808]: I1124 18:22:39.684630 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_dc135c1e-57ee-4be6-835e-b9341e7226e8/openstack-network-exporter/0.log" Nov 24 18:22:39 crc kubenswrapper[4808]: I1124 18:22:39.703706 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_dc135c1e-57ee-4be6-835e-b9341e7226e8/ovsdbserver-nb/0.log" Nov 24 18:22:39 crc kubenswrapper[4808]: I1124 18:22:39.808098 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_67b93485-6ca7-4192-901b-581dbf985068/openstack-network-exporter/0.log" Nov 24 18:22:39 crc kubenswrapper[4808]: I1124 18:22:39.919868 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_67b93485-6ca7-4192-901b-581dbf985068/ovsdbserver-sb/0.log" Nov 24 18:22:39 crc kubenswrapper[4808]: I1124 18:22:39.947850 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-785cbcd98d-bqfnp_18412693-fc60-4860-a2b2-75e830b495a7/placement-api/0.log" Nov 24 18:22:39 crc kubenswrapper[4808]: I1124 18:22:39.996735 4808 generic.go:334] "Generic (PLEG): container finished" podID="c692a6b8-8b25-4a08-83bd-c0567542b35f" containerID="c1b63456b71f0278173a81de37777d66324e12ceaac195739b1b919775a5ac86" exitCode=0 Nov 24 18:22:39 crc kubenswrapper[4808]: I1124 18:22:39.997154 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbqvb" event={"ID":"c692a6b8-8b25-4a08-83bd-c0567542b35f","Type":"ContainerDied","Data":"c1b63456b71f0278173a81de37777d66324e12ceaac195739b1b919775a5ac86"} Nov 24 18:22:40 crc kubenswrapper[4808]: I1124 18:22:40.106784 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-785cbcd98d-bqfnp_18412693-fc60-4860-a2b2-75e830b495a7/placement-log/0.log" Nov 24 18:22:40 crc kubenswrapper[4808]: I1124 18:22:40.140178 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_78791f0f-ea96-45ba-9735-9d10ddebea10/setup-container/0.log" Nov 24 18:22:40 crc kubenswrapper[4808]: I1124 18:22:40.421611 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_78791f0f-ea96-45ba-9735-9d10ddebea10/setup-container/0.log" Nov 24 18:22:40 crc kubenswrapper[4808]: I1124 18:22:40.465043 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_78791f0f-ea96-45ba-9735-9d10ddebea10/rabbitmq/0.log" Nov 24 18:22:40 crc kubenswrapper[4808]: I1124 18:22:40.484288 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bddeab55-cf79-4073-b5c4-420508603e10/setup-container/0.log" Nov 24 18:22:40 crc kubenswrapper[4808]: I1124 18:22:40.632401 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bddeab55-cf79-4073-b5c4-420508603e10/setup-container/0.log" Nov 24 18:22:40 crc kubenswrapper[4808]: I1124 18:22:40.709381 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s_2a9a918d-83b1-4e41-89a1-15e375a5f18e/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:40 crc kubenswrapper[4808]: I1124 18:22:40.756365 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bddeab55-cf79-4073-b5c4-420508603e10/rabbitmq/0.log" Nov 24 18:22:40 crc kubenswrapper[4808]: I1124 18:22:40.905398 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-zkfw7_27ee00de-fd0b-435d-84a5-28cf3c4d0dc7/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:41 crc kubenswrapper[4808]: I1124 18:22:41.030993 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl_a36913c9-4487-4403-9797-8ebacf1304d7/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:41 crc kubenswrapper[4808]: I1124 18:22:41.041202 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbqvb" event={"ID":"c692a6b8-8b25-4a08-83bd-c0567542b35f","Type":"ContainerStarted","Data":"c28af5a3f99840802857b26ad27b6d0697bca72cbdda4a9901ff930739563166"} Nov 24 18:22:41 crc kubenswrapper[4808]: I1124 18:22:41.230123 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-qztqt_92d01f67-8a79-4e09-afad-bf04575f554b/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:41 crc kubenswrapper[4808]: I1124 18:22:41.257992 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-h2xqn_6bab5c2c-7363-40d3-855e-de98d478a482/ssh-known-hosts-edpm-deployment/0.log" Nov 24 18:22:41 crc kubenswrapper[4808]: I1124 18:22:41.538801 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-84b96bd867-hvzp6_a525134b-d25e-4a38-82ec-c7256c6d6697/proxy-server/0.log" Nov 24 18:22:41 crc kubenswrapper[4808]: I1124 18:22:41.553512 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-84b96bd867-hvzp6_a525134b-d25e-4a38-82ec-c7256c6d6697/proxy-httpd/0.log" Nov 24 18:22:41 crc kubenswrapper[4808]: I1124 18:22:41.703654 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-6ttz5_ee5479fc-5360-4541-990d-c8d1a97bae29/swift-ring-rebalance/0.log" Nov 24 18:22:41 crc kubenswrapper[4808]: I1124 18:22:41.738320 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/account-auditor/0.log" Nov 24 18:22:41 crc kubenswrapper[4808]: I1124 18:22:41.777779 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/account-reaper/0.log" Nov 24 18:22:42 crc kubenswrapper[4808]: I1124 18:22:42.118170 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/account-replicator/0.log" Nov 24 18:22:42 crc kubenswrapper[4808]: I1124 18:22:42.166700 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/account-server/0.log" Nov 24 18:22:42 crc kubenswrapper[4808]: I1124 18:22:42.209049 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/container-auditor/0.log" Nov 24 18:22:42 crc kubenswrapper[4808]: I1124 18:22:42.259973 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/container-replicator/0.log" Nov 24 18:22:42 crc kubenswrapper[4808]: I1124 18:22:42.347453 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/container-server/0.log" Nov 24 18:22:42 crc kubenswrapper[4808]: I1124 18:22:42.422580 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/container-updater/0.log" Nov 24 18:22:42 crc kubenswrapper[4808]: I1124 18:22:42.425086 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/object-auditor/0.log" Nov 24 18:22:42 crc kubenswrapper[4808]: I1124 18:22:42.519315 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/object-expirer/0.log" Nov 24 18:22:42 crc kubenswrapper[4808]: I1124 18:22:42.627975 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/object-replicator/0.log" Nov 24 18:22:42 crc kubenswrapper[4808]: I1124 18:22:42.629303 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/object-server/0.log" Nov 24 18:22:42 crc kubenswrapper[4808]: I1124 18:22:42.681053 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/object-updater/0.log" Nov 24 18:22:42 crc kubenswrapper[4808]: I1124 18:22:42.773575 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/rsync/0.log" Nov 24 18:22:42 crc kubenswrapper[4808]: I1124 18:22:42.852521 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/swift-recon-cron/0.log" Nov 24 18:22:42 crc kubenswrapper[4808]: I1124 18:22:42.975677 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-kt77n_7f556606-72e2-4fc7-bb57-cc4bd96c4f9b/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:43 crc kubenswrapper[4808]: I1124 18:22:43.076246 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_d014b1af-f795-43e4-af00-0a152d0cb945/tempest-tests-tempest-tests-runner/0.log" Nov 24 18:22:43 crc kubenswrapper[4808]: I1124 18:22:43.168129 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_5c95e77e-1080-4ca8-96d8-8b34632303ea/test-operator-logs-container/0.log" Nov 24 18:22:43 crc kubenswrapper[4808]: I1124 18:22:43.276746 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-z92sj_579d95f1-e760-4476-bc98-6e33dabf6780/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:22:46 crc kubenswrapper[4808]: I1124 18:22:46.524610 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:46 crc kubenswrapper[4808]: I1124 18:22:46.525220 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:46 crc kubenswrapper[4808]: I1124 18:22:46.576130 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:46 crc kubenswrapper[4808]: I1124 18:22:46.594901 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gbqvb" podStartSLOduration=7.898860265 podStartE2EDuration="10.594883697s" podCreationTimestamp="2025-11-24 18:22:36 +0000 UTC" firstStartedPulling="2025-11-24 18:22:37.951307439 +0000 UTC m=+3350.548975241" lastFinishedPulling="2025-11-24 18:22:40.647330871 +0000 UTC m=+3353.244998673" observedRunningTime="2025-11-24 18:22:41.059344986 +0000 UTC m=+3353.657012778" watchObservedRunningTime="2025-11-24 18:22:46.594883697 +0000 UTC m=+3359.192551499" Nov 24 18:22:47 crc kubenswrapper[4808]: I1124 18:22:47.173337 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:47 crc kubenswrapper[4808]: I1124 18:22:47.257380 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gbqvb"] Nov 24 18:22:49 crc kubenswrapper[4808]: I1124 18:22:49.130522 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gbqvb" podUID="c692a6b8-8b25-4a08-83bd-c0567542b35f" containerName="registry-server" containerID="cri-o://c28af5a3f99840802857b26ad27b6d0697bca72cbdda4a9901ff930739563166" gracePeriod=2 Nov 24 18:22:49 crc kubenswrapper[4808]: I1124 18:22:49.540187 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:49 crc kubenswrapper[4808]: I1124 18:22:49.571262 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c692a6b8-8b25-4a08-83bd-c0567542b35f-utilities\") pod \"c692a6b8-8b25-4a08-83bd-c0567542b35f\" (UID: \"c692a6b8-8b25-4a08-83bd-c0567542b35f\") " Nov 24 18:22:49 crc kubenswrapper[4808]: I1124 18:22:49.571631 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6bcf\" (UniqueName: \"kubernetes.io/projected/c692a6b8-8b25-4a08-83bd-c0567542b35f-kube-api-access-t6bcf\") pod \"c692a6b8-8b25-4a08-83bd-c0567542b35f\" (UID: \"c692a6b8-8b25-4a08-83bd-c0567542b35f\") " Nov 24 18:22:49 crc kubenswrapper[4808]: I1124 18:22:49.571758 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c692a6b8-8b25-4a08-83bd-c0567542b35f-catalog-content\") pod \"c692a6b8-8b25-4a08-83bd-c0567542b35f\" (UID: \"c692a6b8-8b25-4a08-83bd-c0567542b35f\") " Nov 24 18:22:49 crc kubenswrapper[4808]: I1124 18:22:49.577724 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c692a6b8-8b25-4a08-83bd-c0567542b35f-kube-api-access-t6bcf" (OuterVolumeSpecName: "kube-api-access-t6bcf") pod "c692a6b8-8b25-4a08-83bd-c0567542b35f" (UID: "c692a6b8-8b25-4a08-83bd-c0567542b35f"). InnerVolumeSpecName "kube-api-access-t6bcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:22:49 crc kubenswrapper[4808]: I1124 18:22:49.579415 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c692a6b8-8b25-4a08-83bd-c0567542b35f-utilities" (OuterVolumeSpecName: "utilities") pod "c692a6b8-8b25-4a08-83bd-c0567542b35f" (UID: "c692a6b8-8b25-4a08-83bd-c0567542b35f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:22:49 crc kubenswrapper[4808]: I1124 18:22:49.661787 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c692a6b8-8b25-4a08-83bd-c0567542b35f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c692a6b8-8b25-4a08-83bd-c0567542b35f" (UID: "c692a6b8-8b25-4a08-83bd-c0567542b35f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:22:49 crc kubenswrapper[4808]: I1124 18:22:49.677312 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c692a6b8-8b25-4a08-83bd-c0567542b35f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:22:49 crc kubenswrapper[4808]: I1124 18:22:49.677346 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6bcf\" (UniqueName: \"kubernetes.io/projected/c692a6b8-8b25-4a08-83bd-c0567542b35f-kube-api-access-t6bcf\") on node \"crc\" DevicePath \"\"" Nov 24 18:22:49 crc kubenswrapper[4808]: I1124 18:22:49.677358 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c692a6b8-8b25-4a08-83bd-c0567542b35f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:22:49 crc kubenswrapper[4808]: I1124 18:22:49.917499 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_3bd73c68-0bbb-4026-b240-637e5b7c17b5/memcached/0.log" Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.139897 4808 generic.go:334] "Generic (PLEG): container finished" podID="c692a6b8-8b25-4a08-83bd-c0567542b35f" containerID="c28af5a3f99840802857b26ad27b6d0697bca72cbdda4a9901ff930739563166" exitCode=0 Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.139956 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbqvb" event={"ID":"c692a6b8-8b25-4a08-83bd-c0567542b35f","Type":"ContainerDied","Data":"c28af5a3f99840802857b26ad27b6d0697bca72cbdda4a9901ff930739563166"} Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.139965 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbqvb" Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.139989 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbqvb" event={"ID":"c692a6b8-8b25-4a08-83bd-c0567542b35f","Type":"ContainerDied","Data":"1e6150100a2f1bcb780fc2b5e53f9512930e79eefac57ea05e2303a21db7a0a4"} Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.140036 4808 scope.go:117] "RemoveContainer" containerID="c28af5a3f99840802857b26ad27b6d0697bca72cbdda4a9901ff930739563166" Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.164086 4808 scope.go:117] "RemoveContainer" containerID="c1b63456b71f0278173a81de37777d66324e12ceaac195739b1b919775a5ac86" Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.169292 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gbqvb"] Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.177008 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gbqvb"] Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.182695 4808 scope.go:117] "RemoveContainer" containerID="a63399574a962f8f3514ef32013b6e9449aeef3c0cd9532093b8f78546e130ba" Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.235548 4808 scope.go:117] "RemoveContainer" containerID="c28af5a3f99840802857b26ad27b6d0697bca72cbdda4a9901ff930739563166" Nov 24 18:22:50 crc kubenswrapper[4808]: E1124 18:22:50.236883 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c28af5a3f99840802857b26ad27b6d0697bca72cbdda4a9901ff930739563166\": container with ID starting with c28af5a3f99840802857b26ad27b6d0697bca72cbdda4a9901ff930739563166 not found: ID does not exist" containerID="c28af5a3f99840802857b26ad27b6d0697bca72cbdda4a9901ff930739563166" Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.236942 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c28af5a3f99840802857b26ad27b6d0697bca72cbdda4a9901ff930739563166"} err="failed to get container status \"c28af5a3f99840802857b26ad27b6d0697bca72cbdda4a9901ff930739563166\": rpc error: code = NotFound desc = could not find container \"c28af5a3f99840802857b26ad27b6d0697bca72cbdda4a9901ff930739563166\": container with ID starting with c28af5a3f99840802857b26ad27b6d0697bca72cbdda4a9901ff930739563166 not found: ID does not exist" Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.236987 4808 scope.go:117] "RemoveContainer" containerID="c1b63456b71f0278173a81de37777d66324e12ceaac195739b1b919775a5ac86" Nov 24 18:22:50 crc kubenswrapper[4808]: E1124 18:22:50.237373 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1b63456b71f0278173a81de37777d66324e12ceaac195739b1b919775a5ac86\": container with ID starting with c1b63456b71f0278173a81de37777d66324e12ceaac195739b1b919775a5ac86 not found: ID does not exist" containerID="c1b63456b71f0278173a81de37777d66324e12ceaac195739b1b919775a5ac86" Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.237416 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1b63456b71f0278173a81de37777d66324e12ceaac195739b1b919775a5ac86"} err="failed to get container status \"c1b63456b71f0278173a81de37777d66324e12ceaac195739b1b919775a5ac86\": rpc error: code = NotFound desc = could not find container \"c1b63456b71f0278173a81de37777d66324e12ceaac195739b1b919775a5ac86\": container with ID starting with c1b63456b71f0278173a81de37777d66324e12ceaac195739b1b919775a5ac86 not found: ID does not exist" Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.237450 4808 scope.go:117] "RemoveContainer" containerID="a63399574a962f8f3514ef32013b6e9449aeef3c0cd9532093b8f78546e130ba" Nov 24 18:22:50 crc kubenswrapper[4808]: E1124 18:22:50.238091 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a63399574a962f8f3514ef32013b6e9449aeef3c0cd9532093b8f78546e130ba\": container with ID starting with a63399574a962f8f3514ef32013b6e9449aeef3c0cd9532093b8f78546e130ba not found: ID does not exist" containerID="a63399574a962f8f3514ef32013b6e9449aeef3c0cd9532093b8f78546e130ba" Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.238134 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a63399574a962f8f3514ef32013b6e9449aeef3c0cd9532093b8f78546e130ba"} err="failed to get container status \"a63399574a962f8f3514ef32013b6e9449aeef3c0cd9532093b8f78546e130ba\": rpc error: code = NotFound desc = could not find container \"a63399574a962f8f3514ef32013b6e9449aeef3c0cd9532093b8f78546e130ba\": container with ID starting with a63399574a962f8f3514ef32013b6e9449aeef3c0cd9532093b8f78546e130ba not found: ID does not exist" Nov 24 18:22:50 crc kubenswrapper[4808]: I1124 18:22:50.359380 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c692a6b8-8b25-4a08-83bd-c0567542b35f" path="/var/lib/kubelet/pods/c692a6b8-8b25-4a08-83bd-c0567542b35f/volumes" Nov 24 18:23:08 crc kubenswrapper[4808]: I1124 18:23:08.057227 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-bbdb9b8fd-hfwr9_2ece0883-b2fd-4309-b4bd-7639d95445e2/kube-rbac-proxy/0.log" Nov 24 18:23:08 crc kubenswrapper[4808]: I1124 18:23:08.101449 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-bbdb9b8fd-hfwr9_2ece0883-b2fd-4309-b4bd-7639d95445e2/manager/0.log" Nov 24 18:23:08 crc kubenswrapper[4808]: I1124 18:23:08.226626 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-942s7_bf340213-3bb6-4fc3-b8ed-9728393e9a5a/kube-rbac-proxy/0.log" Nov 24 18:23:08 crc kubenswrapper[4808]: I1124 18:23:08.298068 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-942s7_bf340213-3bb6-4fc3-b8ed-9728393e9a5a/manager/0.log" Nov 24 18:23:08 crc kubenswrapper[4808]: I1124 18:23:08.344967 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-nv6j9_edd12c74-ba4f-4cb4-8531-b98607fdab41/kube-rbac-proxy/0.log" Nov 24 18:23:08 crc kubenswrapper[4808]: I1124 18:23:08.473270 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-nv6j9_edd12c74-ba4f-4cb4-8531-b98607fdab41/manager/0.log" Nov 24 18:23:08 crc kubenswrapper[4808]: I1124 18:23:08.489843 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6_eb16e3fe-9175-449c-af90-76b3d2475753/util/0.log" Nov 24 18:23:08 crc kubenswrapper[4808]: I1124 18:23:08.677587 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6_eb16e3fe-9175-449c-af90-76b3d2475753/pull/0.log" Nov 24 18:23:08 crc kubenswrapper[4808]: I1124 18:23:08.693092 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6_eb16e3fe-9175-449c-af90-76b3d2475753/pull/0.log" Nov 24 18:23:08 crc kubenswrapper[4808]: I1124 18:23:08.699733 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6_eb16e3fe-9175-449c-af90-76b3d2475753/util/0.log" Nov 24 18:23:08 crc kubenswrapper[4808]: I1124 18:23:08.838883 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6_eb16e3fe-9175-449c-af90-76b3d2475753/pull/0.log" Nov 24 18:23:08 crc kubenswrapper[4808]: I1124 18:23:08.839814 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6_eb16e3fe-9175-449c-af90-76b3d2475753/util/0.log" Nov 24 18:23:08 crc kubenswrapper[4808]: I1124 18:23:08.893420 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6_eb16e3fe-9175-449c-af90-76b3d2475753/extract/0.log" Nov 24 18:23:09 crc kubenswrapper[4808]: I1124 18:23:09.016662 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7fd7485997-ssvg7_fa1e211d-28e9-4ea1-beb5-6e3d0f5b8b9a/kube-rbac-proxy/0.log" Nov 24 18:23:09 crc kubenswrapper[4808]: I1124 18:23:09.109679 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-758c497787-nnb8j_55559441-f0b0-4353-bbcb-55342cf6ef53/kube-rbac-proxy/0.log" Nov 24 18:23:09 crc kubenswrapper[4808]: I1124 18:23:09.121173 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7fd7485997-ssvg7_fa1e211d-28e9-4ea1-beb5-6e3d0f5b8b9a/manager/0.log" Nov 24 18:23:09 crc kubenswrapper[4808]: I1124 18:23:09.224516 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-758c497787-nnb8j_55559441-f0b0-4353-bbcb-55342cf6ef53/manager/0.log" Nov 24 18:23:09 crc kubenswrapper[4808]: I1124 18:23:09.294761 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-2fzkr_352d8a11-a201-4bc3-8d10-8378a5e04bfb/kube-rbac-proxy/0.log" Nov 24 18:23:09 crc kubenswrapper[4808]: I1124 18:23:09.329366 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-2fzkr_352d8a11-a201-4bc3-8d10-8378a5e04bfb/manager/0.log" Nov 24 18:23:09 crc kubenswrapper[4808]: I1124 18:23:09.456566 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-299sb_58b1610a-0122-447e-a0a0-c2cb927a571f/kube-rbac-proxy/0.log" Nov 24 18:23:09 crc kubenswrapper[4808]: I1124 18:23:09.603325 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-q5xs5_58d593d8-712e-4eb5-8b1c-5df55db3f1d9/kube-rbac-proxy/0.log" Nov 24 18:23:09 crc kubenswrapper[4808]: I1124 18:23:09.688725 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-q5xs5_58d593d8-712e-4eb5-8b1c-5df55db3f1d9/manager/0.log" Nov 24 18:23:09 crc kubenswrapper[4808]: I1124 18:23:09.707879 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-299sb_58b1610a-0122-447e-a0a0-c2cb927a571f/manager/0.log" Nov 24 18:23:09 crc kubenswrapper[4808]: I1124 18:23:09.892988 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-tg4qk_9d1d2914-59b8-4660-80a4-d2a1cc17c2bc/manager/0.log" Nov 24 18:23:09 crc kubenswrapper[4808]: I1124 18:23:09.895486 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-tg4qk_9d1d2914-59b8-4660-80a4-d2a1cc17c2bc/kube-rbac-proxy/0.log" Nov 24 18:23:09 crc kubenswrapper[4808]: I1124 18:23:09.944515 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-8659db65fc-r4hcb_01054a01-64a1-4c81-b03f-5dfdf55ee4cf/kube-rbac-proxy/0.log" Nov 24 18:23:10 crc kubenswrapper[4808]: I1124 18:23:10.062617 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-8659db65fc-r4hcb_01054a01-64a1-4c81-b03f-5dfdf55ee4cf/manager/0.log" Nov 24 18:23:10 crc kubenswrapper[4808]: I1124 18:23:10.119985 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5f598f884b-xqgvh_ac1adbdb-b43a-4d5a-b599-7e2993f72a86/kube-rbac-proxy/0.log" Nov 24 18:23:10 crc kubenswrapper[4808]: I1124 18:23:10.147481 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5f598f884b-xqgvh_ac1adbdb-b43a-4d5a-b599-7e2993f72a86/manager/0.log" Nov 24 18:23:10 crc kubenswrapper[4808]: I1124 18:23:10.301237 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-qrr4f_8a4f09fc-1405-4a74-be87-1b94ee06742a/kube-rbac-proxy/0.log" Nov 24 18:23:10 crc kubenswrapper[4808]: I1124 18:23:10.329118 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-qrr4f_8a4f09fc-1405-4a74-be87-1b94ee06742a/manager/0.log" Nov 24 18:23:10 crc kubenswrapper[4808]: I1124 18:23:10.457919 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-rcwzz_47c249db-b332-44cf-baa5-811e04500c0f/kube-rbac-proxy/0.log" Nov 24 18:23:10 crc kubenswrapper[4808]: I1124 18:23:10.562060 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-b5vdf_af79a806-a6fb-40a1-8ac9-3fccd532c1fc/kube-rbac-proxy/0.log" Nov 24 18:23:10 crc kubenswrapper[4808]: I1124 18:23:10.572751 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-rcwzz_47c249db-b332-44cf-baa5-811e04500c0f/manager/0.log" Nov 24 18:23:10 crc kubenswrapper[4808]: I1124 18:23:10.656347 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-b5vdf_af79a806-a6fb-40a1-8ac9-3fccd532c1fc/manager/0.log" Nov 24 18:23:10 crc kubenswrapper[4808]: I1124 18:23:10.755562 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7d654b7b69f77rn_c796b113-a891-48b2-9a11-50a2577baf7e/manager/0.log" Nov 24 18:23:10 crc kubenswrapper[4808]: I1124 18:23:10.777810 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7d654b7b69f77rn_c796b113-a891-48b2-9a11-50a2577baf7e/kube-rbac-proxy/0.log" Nov 24 18:23:11 crc kubenswrapper[4808]: I1124 18:23:11.080238 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-798b7b4bc9-dhdfm_8d18f465-70ea-4d1f-8531-dcfee37c960d/operator/0.log" Nov 24 18:23:11 crc kubenswrapper[4808]: I1124 18:23:11.223198 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-d6xmw_24eb3123-6595-48d6-8d4a-f984d9a9e601/registry-server/0.log" Nov 24 18:23:11 crc kubenswrapper[4808]: I1124 18:23:11.304848 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-668bd84cc7-lm5c4_dc05be1a-a87d-4829-b2a8-c9d454b7d343/kube-rbac-proxy/0.log" Nov 24 18:23:11 crc kubenswrapper[4808]: I1124 18:23:11.401706 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-668bd84cc7-lm5c4_dc05be1a-a87d-4829-b2a8-c9d454b7d343/manager/0.log" Nov 24 18:23:11 crc kubenswrapper[4808]: I1124 18:23:11.516826 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-zgnpn_321df5bf-732b-4982-841b-3a9a0ddb73f2/kube-rbac-proxy/0.log" Nov 24 18:23:11 crc kubenswrapper[4808]: I1124 18:23:11.547351 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-zgnpn_321df5bf-732b-4982-841b-3a9a0ddb73f2/manager/0.log" Nov 24 18:23:11 crc kubenswrapper[4808]: I1124 18:23:11.703576 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-4r97r_5d6ac280-f3af-4ee8-8933-928ed0c501fc/operator/0.log" Nov 24 18:23:11 crc kubenswrapper[4808]: I1124 18:23:11.841848 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-mdswb_2c56d9f3-3ab5-4f3c-b250-e926807dbfc3/kube-rbac-proxy/0.log" Nov 24 18:23:11 crc kubenswrapper[4808]: I1124 18:23:11.918175 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-mdswb_2c56d9f3-3ab5-4f3c-b250-e926807dbfc3/manager/0.log" Nov 24 18:23:11 crc kubenswrapper[4808]: I1124 18:23:11.975821 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-79c6ffd6b7-vvfcw_e597e7fc-3b56-493b-bcda-10669936636a/manager/0.log" Nov 24 18:23:11 crc kubenswrapper[4808]: I1124 18:23:11.998579 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-9wwt6_ec05f4fb-ca34-4f68-b94e-95b08b21d57b/kube-rbac-proxy/0.log" Nov 24 18:23:12 crc kubenswrapper[4808]: I1124 18:23:12.105206 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-9wwt6_ec05f4fb-ca34-4f68-b94e-95b08b21d57b/manager/0.log" Nov 24 18:23:12 crc kubenswrapper[4808]: I1124 18:23:12.284774 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-66sz4_eb8cf6b6-33aa-427e-9d07-9f95c58c7d30/kube-rbac-proxy/0.log" Nov 24 18:23:12 crc kubenswrapper[4808]: I1124 18:23:12.300662 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-66sz4_eb8cf6b6-33aa-427e-9d07-9f95c58c7d30/manager/0.log" Nov 24 18:23:12 crc kubenswrapper[4808]: I1124 18:23:12.435383 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-wknfh_c62f31fd-04d1-48cc-ae7a-98e1e64c5af7/kube-rbac-proxy/0.log" Nov 24 18:23:12 crc kubenswrapper[4808]: I1124 18:23:12.445683 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-wknfh_c62f31fd-04d1-48cc-ae7a-98e1e64c5af7/manager/0.log" Nov 24 18:23:29 crc kubenswrapper[4808]: I1124 18:23:29.735167 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-ljn6r_4319fe26-e7cc-4665-881a-6275268b6330/control-plane-machine-set-operator/0.log" Nov 24 18:23:29 crc kubenswrapper[4808]: I1124 18:23:29.883276 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qrcr6_d54d24ca-c6c4-4313-8f22-908ce8939026/kube-rbac-proxy/0.log" Nov 24 18:23:29 crc kubenswrapper[4808]: I1124 18:23:29.921755 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qrcr6_d54d24ca-c6c4-4313-8f22-908ce8939026/machine-api-operator/0.log" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.462791 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7l8rm"] Nov 24 18:23:30 crc kubenswrapper[4808]: E1124 18:23:30.463532 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c692a6b8-8b25-4a08-83bd-c0567542b35f" containerName="extract-utilities" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.463563 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c692a6b8-8b25-4a08-83bd-c0567542b35f" containerName="extract-utilities" Nov 24 18:23:30 crc kubenswrapper[4808]: E1124 18:23:30.463603 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c692a6b8-8b25-4a08-83bd-c0567542b35f" containerName="extract-content" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.463617 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c692a6b8-8b25-4a08-83bd-c0567542b35f" containerName="extract-content" Nov 24 18:23:30 crc kubenswrapper[4808]: E1124 18:23:30.463732 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c692a6b8-8b25-4a08-83bd-c0567542b35f" containerName="registry-server" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.463747 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c692a6b8-8b25-4a08-83bd-c0567542b35f" containerName="registry-server" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.464353 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c692a6b8-8b25-4a08-83bd-c0567542b35f" containerName="registry-server" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.466952 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.478008 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7l8rm"] Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.645137 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13ed00f8-b41b-46a5-9231-f92d63189218-utilities\") pod \"certified-operators-7l8rm\" (UID: \"13ed00f8-b41b-46a5-9231-f92d63189218\") " pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.645228 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13ed00f8-b41b-46a5-9231-f92d63189218-catalog-content\") pod \"certified-operators-7l8rm\" (UID: \"13ed00f8-b41b-46a5-9231-f92d63189218\") " pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.645256 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94thg\" (UniqueName: \"kubernetes.io/projected/13ed00f8-b41b-46a5-9231-f92d63189218-kube-api-access-94thg\") pod \"certified-operators-7l8rm\" (UID: \"13ed00f8-b41b-46a5-9231-f92d63189218\") " pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.746590 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13ed00f8-b41b-46a5-9231-f92d63189218-utilities\") pod \"certified-operators-7l8rm\" (UID: \"13ed00f8-b41b-46a5-9231-f92d63189218\") " pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.746683 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13ed00f8-b41b-46a5-9231-f92d63189218-catalog-content\") pod \"certified-operators-7l8rm\" (UID: \"13ed00f8-b41b-46a5-9231-f92d63189218\") " pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.746710 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94thg\" (UniqueName: \"kubernetes.io/projected/13ed00f8-b41b-46a5-9231-f92d63189218-kube-api-access-94thg\") pod \"certified-operators-7l8rm\" (UID: \"13ed00f8-b41b-46a5-9231-f92d63189218\") " pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.747638 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13ed00f8-b41b-46a5-9231-f92d63189218-utilities\") pod \"certified-operators-7l8rm\" (UID: \"13ed00f8-b41b-46a5-9231-f92d63189218\") " pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.747940 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13ed00f8-b41b-46a5-9231-f92d63189218-catalog-content\") pod \"certified-operators-7l8rm\" (UID: \"13ed00f8-b41b-46a5-9231-f92d63189218\") " pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.771982 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94thg\" (UniqueName: \"kubernetes.io/projected/13ed00f8-b41b-46a5-9231-f92d63189218-kube-api-access-94thg\") pod \"certified-operators-7l8rm\" (UID: \"13ed00f8-b41b-46a5-9231-f92d63189218\") " pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:30 crc kubenswrapper[4808]: I1124 18:23:30.800273 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:31 crc kubenswrapper[4808]: I1124 18:23:31.290195 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7l8rm"] Nov 24 18:23:31 crc kubenswrapper[4808]: I1124 18:23:31.557546 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7l8rm" event={"ID":"13ed00f8-b41b-46a5-9231-f92d63189218","Type":"ContainerStarted","Data":"1b42cd62cc08cc34eaa1cef0a093fca731b1a1dfbcf5d07c8c38ae6cbe36d243"} Nov 24 18:23:31 crc kubenswrapper[4808]: I1124 18:23:31.557833 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7l8rm" event={"ID":"13ed00f8-b41b-46a5-9231-f92d63189218","Type":"ContainerStarted","Data":"063226e03dc47362ecaba2e5d4962d23a1132a12acf821d471fb133f1a7180c2"} Nov 24 18:23:32 crc kubenswrapper[4808]: I1124 18:23:32.569996 4808 generic.go:334] "Generic (PLEG): container finished" podID="13ed00f8-b41b-46a5-9231-f92d63189218" containerID="1b42cd62cc08cc34eaa1cef0a093fca731b1a1dfbcf5d07c8c38ae6cbe36d243" exitCode=0 Nov 24 18:23:32 crc kubenswrapper[4808]: I1124 18:23:32.570062 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7l8rm" event={"ID":"13ed00f8-b41b-46a5-9231-f92d63189218","Type":"ContainerDied","Data":"1b42cd62cc08cc34eaa1cef0a093fca731b1a1dfbcf5d07c8c38ae6cbe36d243"} Nov 24 18:23:32 crc kubenswrapper[4808]: I1124 18:23:32.570275 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7l8rm" event={"ID":"13ed00f8-b41b-46a5-9231-f92d63189218","Type":"ContainerStarted","Data":"621338895dc4f207597178f7c7703b0e42d9733dcb90267089617674aa62a745"} Nov 24 18:23:33 crc kubenswrapper[4808]: I1124 18:23:33.586729 4808 generic.go:334] "Generic (PLEG): container finished" podID="13ed00f8-b41b-46a5-9231-f92d63189218" containerID="621338895dc4f207597178f7c7703b0e42d9733dcb90267089617674aa62a745" exitCode=0 Nov 24 18:23:33 crc kubenswrapper[4808]: I1124 18:23:33.586802 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7l8rm" event={"ID":"13ed00f8-b41b-46a5-9231-f92d63189218","Type":"ContainerDied","Data":"621338895dc4f207597178f7c7703b0e42d9733dcb90267089617674aa62a745"} Nov 24 18:23:34 crc kubenswrapper[4808]: I1124 18:23:34.598085 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7l8rm" event={"ID":"13ed00f8-b41b-46a5-9231-f92d63189218","Type":"ContainerStarted","Data":"5c99471bd9db5eb0bb04b742e7593b3d2fa7996a688849a1a675695fa1014845"} Nov 24 18:23:34 crc kubenswrapper[4808]: I1124 18:23:34.621279 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7l8rm" podStartSLOduration=2.18447085 podStartE2EDuration="4.621259962s" podCreationTimestamp="2025-11-24 18:23:30 +0000 UTC" firstStartedPulling="2025-11-24 18:23:31.55904435 +0000 UTC m=+3404.156712152" lastFinishedPulling="2025-11-24 18:23:33.995833462 +0000 UTC m=+3406.593501264" observedRunningTime="2025-11-24 18:23:34.615690276 +0000 UTC m=+3407.213358078" watchObservedRunningTime="2025-11-24 18:23:34.621259962 +0000 UTC m=+3407.218927764" Nov 24 18:23:40 crc kubenswrapper[4808]: I1124 18:23:40.801040 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:40 crc kubenswrapper[4808]: I1124 18:23:40.801646 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:40 crc kubenswrapper[4808]: I1124 18:23:40.870884 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:41 crc kubenswrapper[4808]: I1124 18:23:41.727849 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:41 crc kubenswrapper[4808]: I1124 18:23:41.791735 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7l8rm"] Nov 24 18:23:43 crc kubenswrapper[4808]: I1124 18:23:43.685095 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7l8rm" podUID="13ed00f8-b41b-46a5-9231-f92d63189218" containerName="registry-server" containerID="cri-o://5c99471bd9db5eb0bb04b742e7593b3d2fa7996a688849a1a675695fa1014845" gracePeriod=2 Nov 24 18:23:43 crc kubenswrapper[4808]: I1124 18:23:43.695173 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-2qsx7_016a9f20-1203-4241-9210-55b50eeb60a1/cert-manager-controller/0.log" Nov 24 18:23:43 crc kubenswrapper[4808]: I1124 18:23:43.893917 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-cbqfz_24eadecc-5782-4e5b-8982-46d3e99fa5b1/cert-manager-cainjector/0.log" Nov 24 18:23:43 crc kubenswrapper[4808]: I1124 18:23:43.921205 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-bjlpj_a1ea24f0-1b28-4648-a09e-8c0539792ea1/cert-manager-webhook/0.log" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.132356 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.157921 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94thg\" (UniqueName: \"kubernetes.io/projected/13ed00f8-b41b-46a5-9231-f92d63189218-kube-api-access-94thg\") pod \"13ed00f8-b41b-46a5-9231-f92d63189218\" (UID: \"13ed00f8-b41b-46a5-9231-f92d63189218\") " Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.158006 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13ed00f8-b41b-46a5-9231-f92d63189218-catalog-content\") pod \"13ed00f8-b41b-46a5-9231-f92d63189218\" (UID: \"13ed00f8-b41b-46a5-9231-f92d63189218\") " Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.158082 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13ed00f8-b41b-46a5-9231-f92d63189218-utilities\") pod \"13ed00f8-b41b-46a5-9231-f92d63189218\" (UID: \"13ed00f8-b41b-46a5-9231-f92d63189218\") " Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.159132 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13ed00f8-b41b-46a5-9231-f92d63189218-utilities" (OuterVolumeSpecName: "utilities") pod "13ed00f8-b41b-46a5-9231-f92d63189218" (UID: "13ed00f8-b41b-46a5-9231-f92d63189218"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.163766 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13ed00f8-b41b-46a5-9231-f92d63189218-kube-api-access-94thg" (OuterVolumeSpecName: "kube-api-access-94thg") pod "13ed00f8-b41b-46a5-9231-f92d63189218" (UID: "13ed00f8-b41b-46a5-9231-f92d63189218"). InnerVolumeSpecName "kube-api-access-94thg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.213560 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13ed00f8-b41b-46a5-9231-f92d63189218-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "13ed00f8-b41b-46a5-9231-f92d63189218" (UID: "13ed00f8-b41b-46a5-9231-f92d63189218"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.260832 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13ed00f8-b41b-46a5-9231-f92d63189218-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.260865 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13ed00f8-b41b-46a5-9231-f92d63189218-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.260875 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94thg\" (UniqueName: \"kubernetes.io/projected/13ed00f8-b41b-46a5-9231-f92d63189218-kube-api-access-94thg\") on node \"crc\" DevicePath \"\"" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.701582 4808 generic.go:334] "Generic (PLEG): container finished" podID="13ed00f8-b41b-46a5-9231-f92d63189218" containerID="5c99471bd9db5eb0bb04b742e7593b3d2fa7996a688849a1a675695fa1014845" exitCode=0 Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.701633 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7l8rm" event={"ID":"13ed00f8-b41b-46a5-9231-f92d63189218","Type":"ContainerDied","Data":"5c99471bd9db5eb0bb04b742e7593b3d2fa7996a688849a1a675695fa1014845"} Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.701662 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7l8rm" event={"ID":"13ed00f8-b41b-46a5-9231-f92d63189218","Type":"ContainerDied","Data":"063226e03dc47362ecaba2e5d4962d23a1132a12acf821d471fb133f1a7180c2"} Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.701683 4808 scope.go:117] "RemoveContainer" containerID="5c99471bd9db5eb0bb04b742e7593b3d2fa7996a688849a1a675695fa1014845" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.701828 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7l8rm" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.728879 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7l8rm"] Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.738179 4808 scope.go:117] "RemoveContainer" containerID="621338895dc4f207597178f7c7703b0e42d9733dcb90267089617674aa62a745" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.741663 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7l8rm"] Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.776059 4808 scope.go:117] "RemoveContainer" containerID="1b42cd62cc08cc34eaa1cef0a093fca731b1a1dfbcf5d07c8c38ae6cbe36d243" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.799031 4808 scope.go:117] "RemoveContainer" containerID="5c99471bd9db5eb0bb04b742e7593b3d2fa7996a688849a1a675695fa1014845" Nov 24 18:23:44 crc kubenswrapper[4808]: E1124 18:23:44.799466 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c99471bd9db5eb0bb04b742e7593b3d2fa7996a688849a1a675695fa1014845\": container with ID starting with 5c99471bd9db5eb0bb04b742e7593b3d2fa7996a688849a1a675695fa1014845 not found: ID does not exist" containerID="5c99471bd9db5eb0bb04b742e7593b3d2fa7996a688849a1a675695fa1014845" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.799506 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c99471bd9db5eb0bb04b742e7593b3d2fa7996a688849a1a675695fa1014845"} err="failed to get container status \"5c99471bd9db5eb0bb04b742e7593b3d2fa7996a688849a1a675695fa1014845\": rpc error: code = NotFound desc = could not find container \"5c99471bd9db5eb0bb04b742e7593b3d2fa7996a688849a1a675695fa1014845\": container with ID starting with 5c99471bd9db5eb0bb04b742e7593b3d2fa7996a688849a1a675695fa1014845 not found: ID does not exist" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.799532 4808 scope.go:117] "RemoveContainer" containerID="621338895dc4f207597178f7c7703b0e42d9733dcb90267089617674aa62a745" Nov 24 18:23:44 crc kubenswrapper[4808]: E1124 18:23:44.799765 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"621338895dc4f207597178f7c7703b0e42d9733dcb90267089617674aa62a745\": container with ID starting with 621338895dc4f207597178f7c7703b0e42d9733dcb90267089617674aa62a745 not found: ID does not exist" containerID="621338895dc4f207597178f7c7703b0e42d9733dcb90267089617674aa62a745" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.799788 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"621338895dc4f207597178f7c7703b0e42d9733dcb90267089617674aa62a745"} err="failed to get container status \"621338895dc4f207597178f7c7703b0e42d9733dcb90267089617674aa62a745\": rpc error: code = NotFound desc = could not find container \"621338895dc4f207597178f7c7703b0e42d9733dcb90267089617674aa62a745\": container with ID starting with 621338895dc4f207597178f7c7703b0e42d9733dcb90267089617674aa62a745 not found: ID does not exist" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.799801 4808 scope.go:117] "RemoveContainer" containerID="1b42cd62cc08cc34eaa1cef0a093fca731b1a1dfbcf5d07c8c38ae6cbe36d243" Nov 24 18:23:44 crc kubenswrapper[4808]: E1124 18:23:44.799985 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b42cd62cc08cc34eaa1cef0a093fca731b1a1dfbcf5d07c8c38ae6cbe36d243\": container with ID starting with 1b42cd62cc08cc34eaa1cef0a093fca731b1a1dfbcf5d07c8c38ae6cbe36d243 not found: ID does not exist" containerID="1b42cd62cc08cc34eaa1cef0a093fca731b1a1dfbcf5d07c8c38ae6cbe36d243" Nov 24 18:23:44 crc kubenswrapper[4808]: I1124 18:23:44.800025 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b42cd62cc08cc34eaa1cef0a093fca731b1a1dfbcf5d07c8c38ae6cbe36d243"} err="failed to get container status \"1b42cd62cc08cc34eaa1cef0a093fca731b1a1dfbcf5d07c8c38ae6cbe36d243\": rpc error: code = NotFound desc = could not find container \"1b42cd62cc08cc34eaa1cef0a093fca731b1a1dfbcf5d07c8c38ae6cbe36d243\": container with ID starting with 1b42cd62cc08cc34eaa1cef0a093fca731b1a1dfbcf5d07c8c38ae6cbe36d243 not found: ID does not exist" Nov 24 18:23:46 crc kubenswrapper[4808]: I1124 18:23:46.362705 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13ed00f8-b41b-46a5-9231-f92d63189218" path="/var/lib/kubelet/pods/13ed00f8-b41b-46a5-9231-f92d63189218/volumes" Nov 24 18:23:57 crc kubenswrapper[4808]: I1124 18:23:57.317966 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-4x67c_b9930406-6bd1-4674-a697-e915c209fc08/nmstate-console-plugin/0.log" Nov 24 18:23:57 crc kubenswrapper[4808]: I1124 18:23:57.455633 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-kz6lk_9cc7b294-4f80-47be-88f4-e3b0c010ca1d/nmstate-handler/0.log" Nov 24 18:23:57 crc kubenswrapper[4808]: I1124 18:23:57.490185 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-mf5qh_125495c2-7b26-44c8-913a-0b98c6c81c98/kube-rbac-proxy/0.log" Nov 24 18:23:57 crc kubenswrapper[4808]: I1124 18:23:57.534448 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-mf5qh_125495c2-7b26-44c8-913a-0b98c6c81c98/nmstate-metrics/0.log" Nov 24 18:23:57 crc kubenswrapper[4808]: I1124 18:23:57.679978 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-6dp5q_48267952-ed56-4a20-98c9-e22f04e52c63/nmstate-operator/0.log" Nov 24 18:23:57 crc kubenswrapper[4808]: I1124 18:23:57.715826 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-rzhnm_b1fd2203-60a6-4172-bb4e-124fc185ade8/nmstate-webhook/0.log" Nov 24 18:24:12 crc kubenswrapper[4808]: I1124 18:24:12.972353 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-88twg_9afaca72-2a60-4b1f-a377-4510b24b887e/kube-rbac-proxy/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.096267 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-88twg_9afaca72-2a60-4b1f-a377-4510b24b887e/controller/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.187052 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-frr-files/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.332146 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-reloader/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.364617 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-metrics/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.374059 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-frr-files/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.382221 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-reloader/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.601000 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-metrics/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.601003 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-reloader/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.603549 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-frr-files/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.621330 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-metrics/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.772249 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-reloader/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.774240 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-frr-files/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.782896 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-metrics/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.807403 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/controller/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.978817 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/frr-metrics/0.log" Nov 24 18:24:13 crc kubenswrapper[4808]: I1124 18:24:13.989194 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/kube-rbac-proxy/0.log" Nov 24 18:24:14 crc kubenswrapper[4808]: I1124 18:24:14.011583 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/kube-rbac-proxy-frr/0.log" Nov 24 18:24:14 crc kubenswrapper[4808]: I1124 18:24:14.186221 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/reloader/0.log" Nov 24 18:24:14 crc kubenswrapper[4808]: I1124 18:24:14.243991 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-9qp89_f5d3c911-b65c-46dc-a43a-60247b4d10f3/frr-k8s-webhook-server/0.log" Nov 24 18:24:14 crc kubenswrapper[4808]: I1124 18:24:14.484265 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-56c4766567-nztcn_8dc3fedf-199a-4446-a18a-3dd0f6818f8d/manager/0.log" Nov 24 18:24:14 crc kubenswrapper[4808]: I1124 18:24:14.615170 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7c6ddcbc67-pn9v5_da67a1e6-9fd1-4b27-b965-c907e51b6ce4/webhook-server/0.log" Nov 24 18:24:14 crc kubenswrapper[4808]: I1124 18:24:14.738627 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6mvpd_6f3367d5-54fd-4556-b1ba-5311e15433f3/kube-rbac-proxy/0.log" Nov 24 18:24:15 crc kubenswrapper[4808]: I1124 18:24:15.276522 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/frr/0.log" Nov 24 18:24:15 crc kubenswrapper[4808]: I1124 18:24:15.285141 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6mvpd_6f3367d5-54fd-4556-b1ba-5311e15433f3/speaker/0.log" Nov 24 18:24:28 crc kubenswrapper[4808]: I1124 18:24:28.753423 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp_432d4d4f-23e9-43f7-bf64-af0caa59dbea/util/0.log" Nov 24 18:24:28 crc kubenswrapper[4808]: I1124 18:24:28.919975 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp_432d4d4f-23e9-43f7-bf64-af0caa59dbea/util/0.log" Nov 24 18:24:28 crc kubenswrapper[4808]: I1124 18:24:28.954813 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp_432d4d4f-23e9-43f7-bf64-af0caa59dbea/pull/0.log" Nov 24 18:24:28 crc kubenswrapper[4808]: I1124 18:24:28.987926 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp_432d4d4f-23e9-43f7-bf64-af0caa59dbea/pull/0.log" Nov 24 18:24:29 crc kubenswrapper[4808]: I1124 18:24:29.133456 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp_432d4d4f-23e9-43f7-bf64-af0caa59dbea/pull/0.log" Nov 24 18:24:29 crc kubenswrapper[4808]: I1124 18:24:29.151173 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp_432d4d4f-23e9-43f7-bf64-af0caa59dbea/extract/0.log" Nov 24 18:24:29 crc kubenswrapper[4808]: I1124 18:24:29.159973 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp_432d4d4f-23e9-43f7-bf64-af0caa59dbea/util/0.log" Nov 24 18:24:29 crc kubenswrapper[4808]: I1124 18:24:29.308991 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5j6bd_ba197cbd-9042-410f-adb4-d05cfec94ead/extract-utilities/0.log" Nov 24 18:24:29 crc kubenswrapper[4808]: I1124 18:24:29.500500 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5j6bd_ba197cbd-9042-410f-adb4-d05cfec94ead/extract-utilities/0.log" Nov 24 18:24:29 crc kubenswrapper[4808]: I1124 18:24:29.531900 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5j6bd_ba197cbd-9042-410f-adb4-d05cfec94ead/extract-content/0.log" Nov 24 18:24:29 crc kubenswrapper[4808]: I1124 18:24:29.547167 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5j6bd_ba197cbd-9042-410f-adb4-d05cfec94ead/extract-content/0.log" Nov 24 18:24:29 crc kubenswrapper[4808]: I1124 18:24:29.649102 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5j6bd_ba197cbd-9042-410f-adb4-d05cfec94ead/extract-utilities/0.log" Nov 24 18:24:29 crc kubenswrapper[4808]: I1124 18:24:29.653496 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5j6bd_ba197cbd-9042-410f-adb4-d05cfec94ead/extract-content/0.log" Nov 24 18:24:29 crc kubenswrapper[4808]: I1124 18:24:29.847405 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rbql2_7929b2f8-b7af-4a85-9150-861fe438f602/extract-utilities/0.log" Nov 24 18:24:30 crc kubenswrapper[4808]: I1124 18:24:30.071686 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rbql2_7929b2f8-b7af-4a85-9150-861fe438f602/extract-content/0.log" Nov 24 18:24:30 crc kubenswrapper[4808]: I1124 18:24:30.081518 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rbql2_7929b2f8-b7af-4a85-9150-861fe438f602/extract-content/0.log" Nov 24 18:24:30 crc kubenswrapper[4808]: I1124 18:24:30.081680 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rbql2_7929b2f8-b7af-4a85-9150-861fe438f602/extract-utilities/0.log" Nov 24 18:24:30 crc kubenswrapper[4808]: I1124 18:24:30.411230 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5j6bd_ba197cbd-9042-410f-adb4-d05cfec94ead/registry-server/0.log" Nov 24 18:24:30 crc kubenswrapper[4808]: I1124 18:24:30.462638 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rbql2_7929b2f8-b7af-4a85-9150-861fe438f602/extract-content/0.log" Nov 24 18:24:30 crc kubenswrapper[4808]: I1124 18:24:30.478254 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rbql2_7929b2f8-b7af-4a85-9150-861fe438f602/extract-utilities/0.log" Nov 24 18:24:30 crc kubenswrapper[4808]: I1124 18:24:30.757425 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx_9d683a98-fa75-44d9-8d3d-11552b957864/util/0.log" Nov 24 18:24:30 crc kubenswrapper[4808]: I1124 18:24:30.914677 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx_9d683a98-fa75-44d9-8d3d-11552b957864/pull/0.log" Nov 24 18:24:30 crc kubenswrapper[4808]: I1124 18:24:30.996453 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx_9d683a98-fa75-44d9-8d3d-11552b957864/util/0.log" Nov 24 18:24:31 crc kubenswrapper[4808]: I1124 18:24:31.050370 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx_9d683a98-fa75-44d9-8d3d-11552b957864/pull/0.log" Nov 24 18:24:31 crc kubenswrapper[4808]: I1124 18:24:31.205777 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx_9d683a98-fa75-44d9-8d3d-11552b957864/util/0.log" Nov 24 18:24:31 crc kubenswrapper[4808]: I1124 18:24:31.282693 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx_9d683a98-fa75-44d9-8d3d-11552b957864/pull/0.log" Nov 24 18:24:31 crc kubenswrapper[4808]: I1124 18:24:31.408088 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rbql2_7929b2f8-b7af-4a85-9150-861fe438f602/registry-server/0.log" Nov 24 18:24:31 crc kubenswrapper[4808]: I1124 18:24:31.443920 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx_9d683a98-fa75-44d9-8d3d-11552b957864/extract/0.log" Nov 24 18:24:31 crc kubenswrapper[4808]: I1124 18:24:31.493156 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-555rd_6ae01b29-6123-4219-b106-b8e55e83cb7f/marketplace-operator/0.log" Nov 24 18:24:31 crc kubenswrapper[4808]: I1124 18:24:31.661076 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pxdgf_14cc0e85-6bcb-443d-80f9-5f142532656a/extract-utilities/0.log" Nov 24 18:24:31 crc kubenswrapper[4808]: I1124 18:24:31.788643 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pxdgf_14cc0e85-6bcb-443d-80f9-5f142532656a/extract-utilities/0.log" Nov 24 18:24:31 crc kubenswrapper[4808]: I1124 18:24:31.828316 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pxdgf_14cc0e85-6bcb-443d-80f9-5f142532656a/extract-content/0.log" Nov 24 18:24:31 crc kubenswrapper[4808]: I1124 18:24:31.847609 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pxdgf_14cc0e85-6bcb-443d-80f9-5f142532656a/extract-content/0.log" Nov 24 18:24:32 crc kubenswrapper[4808]: I1124 18:24:32.052353 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pxdgf_14cc0e85-6bcb-443d-80f9-5f142532656a/extract-utilities/0.log" Nov 24 18:24:32 crc kubenswrapper[4808]: I1124 18:24:32.070202 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pxdgf_14cc0e85-6bcb-443d-80f9-5f142532656a/extract-content/0.log" Nov 24 18:24:32 crc kubenswrapper[4808]: I1124 18:24:32.265977 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pxdgf_14cc0e85-6bcb-443d-80f9-5f142532656a/registry-server/0.log" Nov 24 18:24:32 crc kubenswrapper[4808]: I1124 18:24:32.278533 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8kmz9_c13f73b0-c183-4f29-b75d-2fef31356a61/extract-utilities/0.log" Nov 24 18:24:32 crc kubenswrapper[4808]: I1124 18:24:32.461455 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8kmz9_c13f73b0-c183-4f29-b75d-2fef31356a61/extract-utilities/0.log" Nov 24 18:24:32 crc kubenswrapper[4808]: I1124 18:24:32.517257 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8kmz9_c13f73b0-c183-4f29-b75d-2fef31356a61/extract-content/0.log" Nov 24 18:24:32 crc kubenswrapper[4808]: I1124 18:24:32.534225 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8kmz9_c13f73b0-c183-4f29-b75d-2fef31356a61/extract-content/0.log" Nov 24 18:24:32 crc kubenswrapper[4808]: I1124 18:24:32.714667 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8kmz9_c13f73b0-c183-4f29-b75d-2fef31356a61/extract-utilities/0.log" Nov 24 18:24:32 crc kubenswrapper[4808]: I1124 18:24:32.725984 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8kmz9_c13f73b0-c183-4f29-b75d-2fef31356a61/extract-content/0.log" Nov 24 18:24:32 crc kubenswrapper[4808]: I1124 18:24:32.981427 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8kmz9_c13f73b0-c183-4f29-b75d-2fef31356a61/registry-server/0.log" Nov 24 18:24:36 crc kubenswrapper[4808]: I1124 18:24:36.523447 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:24:36 crc kubenswrapper[4808]: I1124 18:24:36.524174 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:25:06 crc kubenswrapper[4808]: I1124 18:25:06.522405 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:25:06 crc kubenswrapper[4808]: I1124 18:25:06.522902 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:25:35 crc kubenswrapper[4808]: I1124 18:25:35.860800 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p8x2r"] Nov 24 18:25:35 crc kubenswrapper[4808]: E1124 18:25:35.862154 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13ed00f8-b41b-46a5-9231-f92d63189218" containerName="registry-server" Nov 24 18:25:35 crc kubenswrapper[4808]: I1124 18:25:35.862176 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="13ed00f8-b41b-46a5-9231-f92d63189218" containerName="registry-server" Nov 24 18:25:35 crc kubenswrapper[4808]: E1124 18:25:35.862217 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13ed00f8-b41b-46a5-9231-f92d63189218" containerName="extract-utilities" Nov 24 18:25:35 crc kubenswrapper[4808]: I1124 18:25:35.862237 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="13ed00f8-b41b-46a5-9231-f92d63189218" containerName="extract-utilities" Nov 24 18:25:35 crc kubenswrapper[4808]: E1124 18:25:35.862274 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13ed00f8-b41b-46a5-9231-f92d63189218" containerName="extract-content" Nov 24 18:25:35 crc kubenswrapper[4808]: I1124 18:25:35.862287 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="13ed00f8-b41b-46a5-9231-f92d63189218" containerName="extract-content" Nov 24 18:25:35 crc kubenswrapper[4808]: I1124 18:25:35.862660 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="13ed00f8-b41b-46a5-9231-f92d63189218" containerName="registry-server" Nov 24 18:25:35 crc kubenswrapper[4808]: I1124 18:25:35.866136 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:35 crc kubenswrapper[4808]: I1124 18:25:35.881761 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8x2r"] Nov 24 18:25:35 crc kubenswrapper[4808]: I1124 18:25:35.996152 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c717c606-03d7-4a58-bc89-d8aeac562846-catalog-content\") pod \"redhat-marketplace-p8x2r\" (UID: \"c717c606-03d7-4a58-bc89-d8aeac562846\") " pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:35 crc kubenswrapper[4808]: I1124 18:25:35.996487 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn2x2\" (UniqueName: \"kubernetes.io/projected/c717c606-03d7-4a58-bc89-d8aeac562846-kube-api-access-bn2x2\") pod \"redhat-marketplace-p8x2r\" (UID: \"c717c606-03d7-4a58-bc89-d8aeac562846\") " pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:35 crc kubenswrapper[4808]: I1124 18:25:35.996638 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c717c606-03d7-4a58-bc89-d8aeac562846-utilities\") pod \"redhat-marketplace-p8x2r\" (UID: \"c717c606-03d7-4a58-bc89-d8aeac562846\") " pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.098307 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn2x2\" (UniqueName: \"kubernetes.io/projected/c717c606-03d7-4a58-bc89-d8aeac562846-kube-api-access-bn2x2\") pod \"redhat-marketplace-p8x2r\" (UID: \"c717c606-03d7-4a58-bc89-d8aeac562846\") " pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.098438 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c717c606-03d7-4a58-bc89-d8aeac562846-utilities\") pod \"redhat-marketplace-p8x2r\" (UID: \"c717c606-03d7-4a58-bc89-d8aeac562846\") " pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.098483 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c717c606-03d7-4a58-bc89-d8aeac562846-catalog-content\") pod \"redhat-marketplace-p8x2r\" (UID: \"c717c606-03d7-4a58-bc89-d8aeac562846\") " pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.099130 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c717c606-03d7-4a58-bc89-d8aeac562846-catalog-content\") pod \"redhat-marketplace-p8x2r\" (UID: \"c717c606-03d7-4a58-bc89-d8aeac562846\") " pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.099256 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c717c606-03d7-4a58-bc89-d8aeac562846-utilities\") pod \"redhat-marketplace-p8x2r\" (UID: \"c717c606-03d7-4a58-bc89-d8aeac562846\") " pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.135073 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn2x2\" (UniqueName: \"kubernetes.io/projected/c717c606-03d7-4a58-bc89-d8aeac562846-kube-api-access-bn2x2\") pod \"redhat-marketplace-p8x2r\" (UID: \"c717c606-03d7-4a58-bc89-d8aeac562846\") " pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.210253 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.522686 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.522976 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.523044 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.524141 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"402d672713f9ca792d3e28c3866699dcd97d837b30faeba8a4bf79ceeb5f226e"} pod="openshift-machine-config-operator/machine-config-daemon-ht87w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.524193 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" containerID="cri-o://402d672713f9ca792d3e28c3866699dcd97d837b30faeba8a4bf79ceeb5f226e" gracePeriod=600 Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.704358 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8x2r"] Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.959160 4808 generic.go:334] "Generic (PLEG): container finished" podID="c717c606-03d7-4a58-bc89-d8aeac562846" containerID="087f15f050e5ba93d38ad39f94b98fd129c48859c1c8b84dd4adfe0ca5d4d5f8" exitCode=0 Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.959253 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8x2r" event={"ID":"c717c606-03d7-4a58-bc89-d8aeac562846","Type":"ContainerDied","Data":"087f15f050e5ba93d38ad39f94b98fd129c48859c1c8b84dd4adfe0ca5d4d5f8"} Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.959347 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8x2r" event={"ID":"c717c606-03d7-4a58-bc89-d8aeac562846","Type":"ContainerStarted","Data":"386f29ea851ba5578fa36e1fa8e7ff44bf14d7f92257897cccad2c8aaa997b90"} Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.977426 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerID="402d672713f9ca792d3e28c3866699dcd97d837b30faeba8a4bf79ceeb5f226e" exitCode=0 Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.977493 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerDied","Data":"402d672713f9ca792d3e28c3866699dcd97d837b30faeba8a4bf79ceeb5f226e"} Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.977537 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395"} Nov 24 18:25:36 crc kubenswrapper[4808]: I1124 18:25:36.977565 4808 scope.go:117] "RemoveContainer" containerID="b3390124641ee207968213367dec77d50d5bb0530b892ebd9ecda8abb3991fff" Nov 24 18:25:38 crc kubenswrapper[4808]: I1124 18:25:38.001972 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8x2r" event={"ID":"c717c606-03d7-4a58-bc89-d8aeac562846","Type":"ContainerStarted","Data":"90fc84445f5bbbf59d8df33a7ea29a2c722363e2d420d08aa83e1b5ad99b0f12"} Nov 24 18:25:39 crc kubenswrapper[4808]: I1124 18:25:39.023749 4808 generic.go:334] "Generic (PLEG): container finished" podID="c717c606-03d7-4a58-bc89-d8aeac562846" containerID="90fc84445f5bbbf59d8df33a7ea29a2c722363e2d420d08aa83e1b5ad99b0f12" exitCode=0 Nov 24 18:25:39 crc kubenswrapper[4808]: I1124 18:25:39.024069 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8x2r" event={"ID":"c717c606-03d7-4a58-bc89-d8aeac562846","Type":"ContainerDied","Data":"90fc84445f5bbbf59d8df33a7ea29a2c722363e2d420d08aa83e1b5ad99b0f12"} Nov 24 18:25:40 crc kubenswrapper[4808]: I1124 18:25:40.044255 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8x2r" event={"ID":"c717c606-03d7-4a58-bc89-d8aeac562846","Type":"ContainerStarted","Data":"fe0d87fa7f066dd0cdf4b859aeb5d76a205fd778776edfc13775be2f1ff251b5"} Nov 24 18:25:40 crc kubenswrapper[4808]: I1124 18:25:40.082633 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p8x2r" podStartSLOduration=2.521406025 podStartE2EDuration="5.08261179s" podCreationTimestamp="2025-11-24 18:25:35 +0000 UTC" firstStartedPulling="2025-11-24 18:25:36.963789385 +0000 UTC m=+3529.561457207" lastFinishedPulling="2025-11-24 18:25:39.52499517 +0000 UTC m=+3532.122662972" observedRunningTime="2025-11-24 18:25:40.082451305 +0000 UTC m=+3532.680119107" watchObservedRunningTime="2025-11-24 18:25:40.08261179 +0000 UTC m=+3532.680279632" Nov 24 18:25:46 crc kubenswrapper[4808]: I1124 18:25:46.210496 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:46 crc kubenswrapper[4808]: I1124 18:25:46.212459 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:46 crc kubenswrapper[4808]: I1124 18:25:46.303861 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:47 crc kubenswrapper[4808]: I1124 18:25:47.192780 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:47 crc kubenswrapper[4808]: I1124 18:25:47.261410 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8x2r"] Nov 24 18:25:49 crc kubenswrapper[4808]: I1124 18:25:49.156615 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p8x2r" podUID="c717c606-03d7-4a58-bc89-d8aeac562846" containerName="registry-server" containerID="cri-o://fe0d87fa7f066dd0cdf4b859aeb5d76a205fd778776edfc13775be2f1ff251b5" gracePeriod=2 Nov 24 18:25:49 crc kubenswrapper[4808]: I1124 18:25:49.777487 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:49 crc kubenswrapper[4808]: I1124 18:25:49.901557 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c717c606-03d7-4a58-bc89-d8aeac562846-utilities\") pod \"c717c606-03d7-4a58-bc89-d8aeac562846\" (UID: \"c717c606-03d7-4a58-bc89-d8aeac562846\") " Nov 24 18:25:49 crc kubenswrapper[4808]: I1124 18:25:49.901713 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c717c606-03d7-4a58-bc89-d8aeac562846-catalog-content\") pod \"c717c606-03d7-4a58-bc89-d8aeac562846\" (UID: \"c717c606-03d7-4a58-bc89-d8aeac562846\") " Nov 24 18:25:49 crc kubenswrapper[4808]: I1124 18:25:49.902045 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bn2x2\" (UniqueName: \"kubernetes.io/projected/c717c606-03d7-4a58-bc89-d8aeac562846-kube-api-access-bn2x2\") pod \"c717c606-03d7-4a58-bc89-d8aeac562846\" (UID: \"c717c606-03d7-4a58-bc89-d8aeac562846\") " Nov 24 18:25:49 crc kubenswrapper[4808]: I1124 18:25:49.902508 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c717c606-03d7-4a58-bc89-d8aeac562846-utilities" (OuterVolumeSpecName: "utilities") pod "c717c606-03d7-4a58-bc89-d8aeac562846" (UID: "c717c606-03d7-4a58-bc89-d8aeac562846"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:25:49 crc kubenswrapper[4808]: I1124 18:25:49.902926 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c717c606-03d7-4a58-bc89-d8aeac562846-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:25:49 crc kubenswrapper[4808]: I1124 18:25:49.909898 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c717c606-03d7-4a58-bc89-d8aeac562846-kube-api-access-bn2x2" (OuterVolumeSpecName: "kube-api-access-bn2x2") pod "c717c606-03d7-4a58-bc89-d8aeac562846" (UID: "c717c606-03d7-4a58-bc89-d8aeac562846"). InnerVolumeSpecName "kube-api-access-bn2x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:25:49 crc kubenswrapper[4808]: I1124 18:25:49.927033 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c717c606-03d7-4a58-bc89-d8aeac562846-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c717c606-03d7-4a58-bc89-d8aeac562846" (UID: "c717c606-03d7-4a58-bc89-d8aeac562846"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.005188 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bn2x2\" (UniqueName: \"kubernetes.io/projected/c717c606-03d7-4a58-bc89-d8aeac562846-kube-api-access-bn2x2\") on node \"crc\" DevicePath \"\"" Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.005221 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c717c606-03d7-4a58-bc89-d8aeac562846-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.167979 4808 generic.go:334] "Generic (PLEG): container finished" podID="c717c606-03d7-4a58-bc89-d8aeac562846" containerID="fe0d87fa7f066dd0cdf4b859aeb5d76a205fd778776edfc13775be2f1ff251b5" exitCode=0 Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.168063 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8x2r" event={"ID":"c717c606-03d7-4a58-bc89-d8aeac562846","Type":"ContainerDied","Data":"fe0d87fa7f066dd0cdf4b859aeb5d76a205fd778776edfc13775be2f1ff251b5"} Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.168090 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8x2r" Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.168120 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8x2r" event={"ID":"c717c606-03d7-4a58-bc89-d8aeac562846","Type":"ContainerDied","Data":"386f29ea851ba5578fa36e1fa8e7ff44bf14d7f92257897cccad2c8aaa997b90"} Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.168143 4808 scope.go:117] "RemoveContainer" containerID="fe0d87fa7f066dd0cdf4b859aeb5d76a205fd778776edfc13775be2f1ff251b5" Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.193848 4808 scope.go:117] "RemoveContainer" containerID="90fc84445f5bbbf59d8df33a7ea29a2c722363e2d420d08aa83e1b5ad99b0f12" Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.216722 4808 scope.go:117] "RemoveContainer" containerID="087f15f050e5ba93d38ad39f94b98fd129c48859c1c8b84dd4adfe0ca5d4d5f8" Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.226010 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8x2r"] Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.235511 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8x2r"] Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.274867 4808 scope.go:117] "RemoveContainer" containerID="fe0d87fa7f066dd0cdf4b859aeb5d76a205fd778776edfc13775be2f1ff251b5" Nov 24 18:25:50 crc kubenswrapper[4808]: E1124 18:25:50.275325 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe0d87fa7f066dd0cdf4b859aeb5d76a205fd778776edfc13775be2f1ff251b5\": container with ID starting with fe0d87fa7f066dd0cdf4b859aeb5d76a205fd778776edfc13775be2f1ff251b5 not found: ID does not exist" containerID="fe0d87fa7f066dd0cdf4b859aeb5d76a205fd778776edfc13775be2f1ff251b5" Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.275364 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe0d87fa7f066dd0cdf4b859aeb5d76a205fd778776edfc13775be2f1ff251b5"} err="failed to get container status \"fe0d87fa7f066dd0cdf4b859aeb5d76a205fd778776edfc13775be2f1ff251b5\": rpc error: code = NotFound desc = could not find container \"fe0d87fa7f066dd0cdf4b859aeb5d76a205fd778776edfc13775be2f1ff251b5\": container with ID starting with fe0d87fa7f066dd0cdf4b859aeb5d76a205fd778776edfc13775be2f1ff251b5 not found: ID does not exist" Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.275389 4808 scope.go:117] "RemoveContainer" containerID="90fc84445f5bbbf59d8df33a7ea29a2c722363e2d420d08aa83e1b5ad99b0f12" Nov 24 18:25:50 crc kubenswrapper[4808]: E1124 18:25:50.276160 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90fc84445f5bbbf59d8df33a7ea29a2c722363e2d420d08aa83e1b5ad99b0f12\": container with ID starting with 90fc84445f5bbbf59d8df33a7ea29a2c722363e2d420d08aa83e1b5ad99b0f12 not found: ID does not exist" containerID="90fc84445f5bbbf59d8df33a7ea29a2c722363e2d420d08aa83e1b5ad99b0f12" Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.276191 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90fc84445f5bbbf59d8df33a7ea29a2c722363e2d420d08aa83e1b5ad99b0f12"} err="failed to get container status \"90fc84445f5bbbf59d8df33a7ea29a2c722363e2d420d08aa83e1b5ad99b0f12\": rpc error: code = NotFound desc = could not find container \"90fc84445f5bbbf59d8df33a7ea29a2c722363e2d420d08aa83e1b5ad99b0f12\": container with ID starting with 90fc84445f5bbbf59d8df33a7ea29a2c722363e2d420d08aa83e1b5ad99b0f12 not found: ID does not exist" Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.276224 4808 scope.go:117] "RemoveContainer" containerID="087f15f050e5ba93d38ad39f94b98fd129c48859c1c8b84dd4adfe0ca5d4d5f8" Nov 24 18:25:50 crc kubenswrapper[4808]: E1124 18:25:50.276525 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"087f15f050e5ba93d38ad39f94b98fd129c48859c1c8b84dd4adfe0ca5d4d5f8\": container with ID starting with 087f15f050e5ba93d38ad39f94b98fd129c48859c1c8b84dd4adfe0ca5d4d5f8 not found: ID does not exist" containerID="087f15f050e5ba93d38ad39f94b98fd129c48859c1c8b84dd4adfe0ca5d4d5f8" Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.276554 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"087f15f050e5ba93d38ad39f94b98fd129c48859c1c8b84dd4adfe0ca5d4d5f8"} err="failed to get container status \"087f15f050e5ba93d38ad39f94b98fd129c48859c1c8b84dd4adfe0ca5d4d5f8\": rpc error: code = NotFound desc = could not find container \"087f15f050e5ba93d38ad39f94b98fd129c48859c1c8b84dd4adfe0ca5d4d5f8\": container with ID starting with 087f15f050e5ba93d38ad39f94b98fd129c48859c1c8b84dd4adfe0ca5d4d5f8 not found: ID does not exist" Nov 24 18:25:50 crc kubenswrapper[4808]: I1124 18:25:50.368647 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c717c606-03d7-4a58-bc89-d8aeac562846" path="/var/lib/kubelet/pods/c717c606-03d7-4a58-bc89-d8aeac562846/volumes" Nov 24 18:26:09 crc kubenswrapper[4808]: I1124 18:26:09.403921 4808 generic.go:334] "Generic (PLEG): container finished" podID="b3d20d4f-5b20-4f82-8e89-f232fd4abfbe" containerID="e71fec119fe999e3b12414f1e1f0b907fde36af74c69b188eab679872232a824" exitCode=0 Nov 24 18:26:09 crc kubenswrapper[4808]: I1124 18:26:09.404101 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bxz78/must-gather-tbb5f" event={"ID":"b3d20d4f-5b20-4f82-8e89-f232fd4abfbe","Type":"ContainerDied","Data":"e71fec119fe999e3b12414f1e1f0b907fde36af74c69b188eab679872232a824"} Nov 24 18:26:09 crc kubenswrapper[4808]: I1124 18:26:09.405769 4808 scope.go:117] "RemoveContainer" containerID="e71fec119fe999e3b12414f1e1f0b907fde36af74c69b188eab679872232a824" Nov 24 18:26:09 crc kubenswrapper[4808]: I1124 18:26:09.536063 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bxz78_must-gather-tbb5f_b3d20d4f-5b20-4f82-8e89-f232fd4abfbe/gather/0.log" Nov 24 18:26:17 crc kubenswrapper[4808]: I1124 18:26:17.153804 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bxz78/must-gather-tbb5f"] Nov 24 18:26:17 crc kubenswrapper[4808]: I1124 18:26:17.154817 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-bxz78/must-gather-tbb5f" podUID="b3d20d4f-5b20-4f82-8e89-f232fd4abfbe" containerName="copy" containerID="cri-o://70ecfba37d6166ed8dc67a6787df0f1d522a6b092bd4902f4fb542fc1bd46c09" gracePeriod=2 Nov 24 18:26:17 crc kubenswrapper[4808]: I1124 18:26:17.165059 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bxz78/must-gather-tbb5f"] Nov 24 18:26:17 crc kubenswrapper[4808]: I1124 18:26:17.507378 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bxz78_must-gather-tbb5f_b3d20d4f-5b20-4f82-8e89-f232fd4abfbe/copy/0.log" Nov 24 18:26:17 crc kubenswrapper[4808]: I1124 18:26:17.508250 4808 generic.go:334] "Generic (PLEG): container finished" podID="b3d20d4f-5b20-4f82-8e89-f232fd4abfbe" containerID="70ecfba37d6166ed8dc67a6787df0f1d522a6b092bd4902f4fb542fc1bd46c09" exitCode=143 Nov 24 18:26:17 crc kubenswrapper[4808]: I1124 18:26:17.652605 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bxz78_must-gather-tbb5f_b3d20d4f-5b20-4f82-8e89-f232fd4abfbe/copy/0.log" Nov 24 18:26:17 crc kubenswrapper[4808]: I1124 18:26:17.653037 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/must-gather-tbb5f" Nov 24 18:26:17 crc kubenswrapper[4808]: I1124 18:26:17.730306 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stfkb\" (UniqueName: \"kubernetes.io/projected/b3d20d4f-5b20-4f82-8e89-f232fd4abfbe-kube-api-access-stfkb\") pod \"b3d20d4f-5b20-4f82-8e89-f232fd4abfbe\" (UID: \"b3d20d4f-5b20-4f82-8e89-f232fd4abfbe\") " Nov 24 18:26:17 crc kubenswrapper[4808]: I1124 18:26:17.730440 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b3d20d4f-5b20-4f82-8e89-f232fd4abfbe-must-gather-output\") pod \"b3d20d4f-5b20-4f82-8e89-f232fd4abfbe\" (UID: \"b3d20d4f-5b20-4f82-8e89-f232fd4abfbe\") " Nov 24 18:26:17 crc kubenswrapper[4808]: I1124 18:26:17.741444 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3d20d4f-5b20-4f82-8e89-f232fd4abfbe-kube-api-access-stfkb" (OuterVolumeSpecName: "kube-api-access-stfkb") pod "b3d20d4f-5b20-4f82-8e89-f232fd4abfbe" (UID: "b3d20d4f-5b20-4f82-8e89-f232fd4abfbe"). InnerVolumeSpecName "kube-api-access-stfkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:26:17 crc kubenswrapper[4808]: I1124 18:26:17.832948 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stfkb\" (UniqueName: \"kubernetes.io/projected/b3d20d4f-5b20-4f82-8e89-f232fd4abfbe-kube-api-access-stfkb\") on node \"crc\" DevicePath \"\"" Nov 24 18:26:17 crc kubenswrapper[4808]: I1124 18:26:17.888534 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3d20d4f-5b20-4f82-8e89-f232fd4abfbe-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b3d20d4f-5b20-4f82-8e89-f232fd4abfbe" (UID: "b3d20d4f-5b20-4f82-8e89-f232fd4abfbe"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:26:17 crc kubenswrapper[4808]: I1124 18:26:17.935407 4808 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b3d20d4f-5b20-4f82-8e89-f232fd4abfbe-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 18:26:18 crc kubenswrapper[4808]: I1124 18:26:18.356952 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3d20d4f-5b20-4f82-8e89-f232fd4abfbe" path="/var/lib/kubelet/pods/b3d20d4f-5b20-4f82-8e89-f232fd4abfbe/volumes" Nov 24 18:26:18 crc kubenswrapper[4808]: I1124 18:26:18.520505 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bxz78_must-gather-tbb5f_b3d20d4f-5b20-4f82-8e89-f232fd4abfbe/copy/0.log" Nov 24 18:26:18 crc kubenswrapper[4808]: I1124 18:26:18.520928 4808 scope.go:117] "RemoveContainer" containerID="70ecfba37d6166ed8dc67a6787df0f1d522a6b092bd4902f4fb542fc1bd46c09" Nov 24 18:26:18 crc kubenswrapper[4808]: I1124 18:26:18.521203 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bxz78/must-gather-tbb5f" Nov 24 18:26:18 crc kubenswrapper[4808]: I1124 18:26:18.543123 4808 scope.go:117] "RemoveContainer" containerID="e71fec119fe999e3b12414f1e1f0b907fde36af74c69b188eab679872232a824" Nov 24 18:27:36 crc kubenswrapper[4808]: I1124 18:27:36.522533 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:27:36 crc kubenswrapper[4808]: I1124 18:27:36.523578 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:27:55 crc kubenswrapper[4808]: I1124 18:27:55.218256 4808 scope.go:117] "RemoveContainer" containerID="2f7ef1f21801221f023f341824b577cbe61fb18a26a475ad4a1cbedcfe60e0f5" Nov 24 18:28:06 crc kubenswrapper[4808]: I1124 18:28:06.522987 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:28:06 crc kubenswrapper[4808]: I1124 18:28:06.523661 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:28:36 crc kubenswrapper[4808]: I1124 18:28:36.523304 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:28:36 crc kubenswrapper[4808]: I1124 18:28:36.523838 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:28:36 crc kubenswrapper[4808]: I1124 18:28:36.523882 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 18:28:36 crc kubenswrapper[4808]: I1124 18:28:36.524526 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395"} pod="openshift-machine-config-operator/machine-config-daemon-ht87w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 18:28:36 crc kubenswrapper[4808]: I1124 18:28:36.524575 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" containerID="cri-o://65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" gracePeriod=600 Nov 24 18:28:36 crc kubenswrapper[4808]: E1124 18:28:36.648363 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:28:37 crc kubenswrapper[4808]: I1124 18:28:37.555986 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" exitCode=0 Nov 24 18:28:37 crc kubenswrapper[4808]: I1124 18:28:37.556061 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerDied","Data":"65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395"} Nov 24 18:28:37 crc kubenswrapper[4808]: I1124 18:28:37.556320 4808 scope.go:117] "RemoveContainer" containerID="402d672713f9ca792d3e28c3866699dcd97d837b30faeba8a4bf79ceeb5f226e" Nov 24 18:28:37 crc kubenswrapper[4808]: I1124 18:28:37.557347 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:28:37 crc kubenswrapper[4808]: E1124 18:28:37.557938 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:28:49 crc kubenswrapper[4808]: I1124 18:28:49.348250 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:28:49 crc kubenswrapper[4808]: E1124 18:28:49.349207 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:29:01 crc kubenswrapper[4808]: I1124 18:29:01.347769 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:29:01 crc kubenswrapper[4808]: E1124 18:29:01.348865 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:29:01 crc kubenswrapper[4808]: I1124 18:29:01.989301 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-d2lp6/must-gather-hjrb8"] Nov 24 18:29:01 crc kubenswrapper[4808]: E1124 18:29:01.990142 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d20d4f-5b20-4f82-8e89-f232fd4abfbe" containerName="copy" Nov 24 18:29:01 crc kubenswrapper[4808]: I1124 18:29:01.990165 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d20d4f-5b20-4f82-8e89-f232fd4abfbe" containerName="copy" Nov 24 18:29:01 crc kubenswrapper[4808]: E1124 18:29:01.990189 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d20d4f-5b20-4f82-8e89-f232fd4abfbe" containerName="gather" Nov 24 18:29:01 crc kubenswrapper[4808]: I1124 18:29:01.990199 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d20d4f-5b20-4f82-8e89-f232fd4abfbe" containerName="gather" Nov 24 18:29:01 crc kubenswrapper[4808]: E1124 18:29:01.990236 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c717c606-03d7-4a58-bc89-d8aeac562846" containerName="registry-server" Nov 24 18:29:01 crc kubenswrapper[4808]: I1124 18:29:01.990244 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c717c606-03d7-4a58-bc89-d8aeac562846" containerName="registry-server" Nov 24 18:29:01 crc kubenswrapper[4808]: E1124 18:29:01.990260 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c717c606-03d7-4a58-bc89-d8aeac562846" containerName="extract-content" Nov 24 18:29:01 crc kubenswrapper[4808]: I1124 18:29:01.990267 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c717c606-03d7-4a58-bc89-d8aeac562846" containerName="extract-content" Nov 24 18:29:01 crc kubenswrapper[4808]: E1124 18:29:01.990286 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c717c606-03d7-4a58-bc89-d8aeac562846" containerName="extract-utilities" Nov 24 18:29:01 crc kubenswrapper[4808]: I1124 18:29:01.990294 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c717c606-03d7-4a58-bc89-d8aeac562846" containerName="extract-utilities" Nov 24 18:29:01 crc kubenswrapper[4808]: I1124 18:29:01.990499 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3d20d4f-5b20-4f82-8e89-f232fd4abfbe" containerName="copy" Nov 24 18:29:01 crc kubenswrapper[4808]: I1124 18:29:01.990518 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c717c606-03d7-4a58-bc89-d8aeac562846" containerName="registry-server" Nov 24 18:29:01 crc kubenswrapper[4808]: I1124 18:29:01.990545 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3d20d4f-5b20-4f82-8e89-f232fd4abfbe" containerName="gather" Nov 24 18:29:01 crc kubenswrapper[4808]: I1124 18:29:01.991683 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/must-gather-hjrb8" Nov 24 18:29:01 crc kubenswrapper[4808]: I1124 18:29:01.995510 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-d2lp6"/"default-dockercfg-4n9nm" Nov 24 18:29:01 crc kubenswrapper[4808]: I1124 18:29:01.995832 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-d2lp6"/"kube-root-ca.crt" Nov 24 18:29:01 crc kubenswrapper[4808]: I1124 18:29:01.999667 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-d2lp6"/"openshift-service-ca.crt" Nov 24 18:29:02 crc kubenswrapper[4808]: I1124 18:29:02.023079 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-d2lp6/must-gather-hjrb8"] Nov 24 18:29:02 crc kubenswrapper[4808]: I1124 18:29:02.125540 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88rdg\" (UniqueName: \"kubernetes.io/projected/626e8db0-9921-43ac-a1e1-55ff96937a43-kube-api-access-88rdg\") pod \"must-gather-hjrb8\" (UID: \"626e8db0-9921-43ac-a1e1-55ff96937a43\") " pod="openshift-must-gather-d2lp6/must-gather-hjrb8" Nov 24 18:29:02 crc kubenswrapper[4808]: I1124 18:29:02.125625 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/626e8db0-9921-43ac-a1e1-55ff96937a43-must-gather-output\") pod \"must-gather-hjrb8\" (UID: \"626e8db0-9921-43ac-a1e1-55ff96937a43\") " pod="openshift-must-gather-d2lp6/must-gather-hjrb8" Nov 24 18:29:02 crc kubenswrapper[4808]: I1124 18:29:02.227622 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88rdg\" (UniqueName: \"kubernetes.io/projected/626e8db0-9921-43ac-a1e1-55ff96937a43-kube-api-access-88rdg\") pod \"must-gather-hjrb8\" (UID: \"626e8db0-9921-43ac-a1e1-55ff96937a43\") " pod="openshift-must-gather-d2lp6/must-gather-hjrb8" Nov 24 18:29:02 crc kubenswrapper[4808]: I1124 18:29:02.227700 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/626e8db0-9921-43ac-a1e1-55ff96937a43-must-gather-output\") pod \"must-gather-hjrb8\" (UID: \"626e8db0-9921-43ac-a1e1-55ff96937a43\") " pod="openshift-must-gather-d2lp6/must-gather-hjrb8" Nov 24 18:29:02 crc kubenswrapper[4808]: I1124 18:29:02.228408 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/626e8db0-9921-43ac-a1e1-55ff96937a43-must-gather-output\") pod \"must-gather-hjrb8\" (UID: \"626e8db0-9921-43ac-a1e1-55ff96937a43\") " pod="openshift-must-gather-d2lp6/must-gather-hjrb8" Nov 24 18:29:02 crc kubenswrapper[4808]: I1124 18:29:02.257578 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88rdg\" (UniqueName: \"kubernetes.io/projected/626e8db0-9921-43ac-a1e1-55ff96937a43-kube-api-access-88rdg\") pod \"must-gather-hjrb8\" (UID: \"626e8db0-9921-43ac-a1e1-55ff96937a43\") " pod="openshift-must-gather-d2lp6/must-gather-hjrb8" Nov 24 18:29:02 crc kubenswrapper[4808]: I1124 18:29:02.323538 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/must-gather-hjrb8" Nov 24 18:29:02 crc kubenswrapper[4808]: I1124 18:29:02.595136 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-d2lp6/must-gather-hjrb8"] Nov 24 18:29:02 crc kubenswrapper[4808]: I1124 18:29:02.845876 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2lp6/must-gather-hjrb8" event={"ID":"626e8db0-9921-43ac-a1e1-55ff96937a43","Type":"ContainerStarted","Data":"db4a7c6b2cb3d4afd69ae23ecd329060a4b15e1b6c44c79769b31266f09ee54a"} Nov 24 18:29:03 crc kubenswrapper[4808]: I1124 18:29:03.868292 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2lp6/must-gather-hjrb8" event={"ID":"626e8db0-9921-43ac-a1e1-55ff96937a43","Type":"ContainerStarted","Data":"f90518d8aa5aad6b9761866d87910ff71bac9813edb2c684c3570b4efe803eb1"} Nov 24 18:29:03 crc kubenswrapper[4808]: I1124 18:29:03.868343 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2lp6/must-gather-hjrb8" event={"ID":"626e8db0-9921-43ac-a1e1-55ff96937a43","Type":"ContainerStarted","Data":"71bdc92f32291232187da14655de079d7bce31d4f354b1c4ee44415bbcba1bf2"} Nov 24 18:29:03 crc kubenswrapper[4808]: I1124 18:29:03.889534 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-d2lp6/must-gather-hjrb8" podStartSLOduration=2.889515994 podStartE2EDuration="2.889515994s" podCreationTimestamp="2025-11-24 18:29:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 18:29:03.887930849 +0000 UTC m=+3736.485598651" watchObservedRunningTime="2025-11-24 18:29:03.889515994 +0000 UTC m=+3736.487183796" Nov 24 18:29:06 crc kubenswrapper[4808]: I1124 18:29:06.526567 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-d2lp6/crc-debug-gxxrg"] Nov 24 18:29:06 crc kubenswrapper[4808]: I1124 18:29:06.528313 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/crc-debug-gxxrg" Nov 24 18:29:06 crc kubenswrapper[4808]: I1124 18:29:06.629101 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/64482799-287e-4dea-b802-20ba3b468319-host\") pod \"crc-debug-gxxrg\" (UID: \"64482799-287e-4dea-b802-20ba3b468319\") " pod="openshift-must-gather-d2lp6/crc-debug-gxxrg" Nov 24 18:29:06 crc kubenswrapper[4808]: I1124 18:29:06.629266 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vk5z\" (UniqueName: \"kubernetes.io/projected/64482799-287e-4dea-b802-20ba3b468319-kube-api-access-6vk5z\") pod \"crc-debug-gxxrg\" (UID: \"64482799-287e-4dea-b802-20ba3b468319\") " pod="openshift-must-gather-d2lp6/crc-debug-gxxrg" Nov 24 18:29:06 crc kubenswrapper[4808]: I1124 18:29:06.730839 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vk5z\" (UniqueName: \"kubernetes.io/projected/64482799-287e-4dea-b802-20ba3b468319-kube-api-access-6vk5z\") pod \"crc-debug-gxxrg\" (UID: \"64482799-287e-4dea-b802-20ba3b468319\") " pod="openshift-must-gather-d2lp6/crc-debug-gxxrg" Nov 24 18:29:06 crc kubenswrapper[4808]: I1124 18:29:06.730965 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/64482799-287e-4dea-b802-20ba3b468319-host\") pod \"crc-debug-gxxrg\" (UID: \"64482799-287e-4dea-b802-20ba3b468319\") " pod="openshift-must-gather-d2lp6/crc-debug-gxxrg" Nov 24 18:29:06 crc kubenswrapper[4808]: I1124 18:29:06.731165 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/64482799-287e-4dea-b802-20ba3b468319-host\") pod \"crc-debug-gxxrg\" (UID: \"64482799-287e-4dea-b802-20ba3b468319\") " pod="openshift-must-gather-d2lp6/crc-debug-gxxrg" Nov 24 18:29:06 crc kubenswrapper[4808]: I1124 18:29:06.751139 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vk5z\" (UniqueName: \"kubernetes.io/projected/64482799-287e-4dea-b802-20ba3b468319-kube-api-access-6vk5z\") pod \"crc-debug-gxxrg\" (UID: \"64482799-287e-4dea-b802-20ba3b468319\") " pod="openshift-must-gather-d2lp6/crc-debug-gxxrg" Nov 24 18:29:06 crc kubenswrapper[4808]: I1124 18:29:06.857646 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/crc-debug-gxxrg" Nov 24 18:29:06 crc kubenswrapper[4808]: W1124 18:29:06.890604 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64482799_287e_4dea_b802_20ba3b468319.slice/crio-b64ed14fe773da77e980a306ce0e524232b0a4d1f5816e610de5a0f9b4920638 WatchSource:0}: Error finding container b64ed14fe773da77e980a306ce0e524232b0a4d1f5816e610de5a0f9b4920638: Status 404 returned error can't find the container with id b64ed14fe773da77e980a306ce0e524232b0a4d1f5816e610de5a0f9b4920638 Nov 24 18:29:07 crc kubenswrapper[4808]: I1124 18:29:07.905923 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2lp6/crc-debug-gxxrg" event={"ID":"64482799-287e-4dea-b802-20ba3b468319","Type":"ContainerStarted","Data":"39518ab101eb2177631df8a3e019cbb9530fd34f620b752bd70e5791befab7ae"} Nov 24 18:29:07 crc kubenswrapper[4808]: I1124 18:29:07.906519 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2lp6/crc-debug-gxxrg" event={"ID":"64482799-287e-4dea-b802-20ba3b468319","Type":"ContainerStarted","Data":"b64ed14fe773da77e980a306ce0e524232b0a4d1f5816e610de5a0f9b4920638"} Nov 24 18:29:07 crc kubenswrapper[4808]: I1124 18:29:07.949526 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-d2lp6/crc-debug-gxxrg" podStartSLOduration=1.949493866 podStartE2EDuration="1.949493866s" podCreationTimestamp="2025-11-24 18:29:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 18:29:07.924714772 +0000 UTC m=+3740.522382564" watchObservedRunningTime="2025-11-24 18:29:07.949493866 +0000 UTC m=+3740.547161678" Nov 24 18:29:16 crc kubenswrapper[4808]: I1124 18:29:16.346915 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:29:16 crc kubenswrapper[4808]: E1124 18:29:16.347644 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:29:31 crc kubenswrapper[4808]: I1124 18:29:31.347475 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:29:31 crc kubenswrapper[4808]: E1124 18:29:31.348385 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:29:40 crc kubenswrapper[4808]: I1124 18:29:40.218543 4808 generic.go:334] "Generic (PLEG): container finished" podID="64482799-287e-4dea-b802-20ba3b468319" containerID="39518ab101eb2177631df8a3e019cbb9530fd34f620b752bd70e5791befab7ae" exitCode=0 Nov 24 18:29:40 crc kubenswrapper[4808]: I1124 18:29:40.218631 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2lp6/crc-debug-gxxrg" event={"ID":"64482799-287e-4dea-b802-20ba3b468319","Type":"ContainerDied","Data":"39518ab101eb2177631df8a3e019cbb9530fd34f620b752bd70e5791befab7ae"} Nov 24 18:29:41 crc kubenswrapper[4808]: I1124 18:29:41.321341 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/crc-debug-gxxrg" Nov 24 18:29:41 crc kubenswrapper[4808]: I1124 18:29:41.352692 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-d2lp6/crc-debug-gxxrg"] Nov 24 18:29:41 crc kubenswrapper[4808]: I1124 18:29:41.364403 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-d2lp6/crc-debug-gxxrg"] Nov 24 18:29:41 crc kubenswrapper[4808]: I1124 18:29:41.463696 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vk5z\" (UniqueName: \"kubernetes.io/projected/64482799-287e-4dea-b802-20ba3b468319-kube-api-access-6vk5z\") pod \"64482799-287e-4dea-b802-20ba3b468319\" (UID: \"64482799-287e-4dea-b802-20ba3b468319\") " Nov 24 18:29:41 crc kubenswrapper[4808]: I1124 18:29:41.464115 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/64482799-287e-4dea-b802-20ba3b468319-host\") pod \"64482799-287e-4dea-b802-20ba3b468319\" (UID: \"64482799-287e-4dea-b802-20ba3b468319\") " Nov 24 18:29:41 crc kubenswrapper[4808]: I1124 18:29:41.464178 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/64482799-287e-4dea-b802-20ba3b468319-host" (OuterVolumeSpecName: "host") pod "64482799-287e-4dea-b802-20ba3b468319" (UID: "64482799-287e-4dea-b802-20ba3b468319"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 18:29:41 crc kubenswrapper[4808]: I1124 18:29:41.464572 4808 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/64482799-287e-4dea-b802-20ba3b468319-host\") on node \"crc\" DevicePath \"\"" Nov 24 18:29:41 crc kubenswrapper[4808]: I1124 18:29:41.482707 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64482799-287e-4dea-b802-20ba3b468319-kube-api-access-6vk5z" (OuterVolumeSpecName: "kube-api-access-6vk5z") pod "64482799-287e-4dea-b802-20ba3b468319" (UID: "64482799-287e-4dea-b802-20ba3b468319"). InnerVolumeSpecName "kube-api-access-6vk5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:29:41 crc kubenswrapper[4808]: I1124 18:29:41.566624 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vk5z\" (UniqueName: \"kubernetes.io/projected/64482799-287e-4dea-b802-20ba3b468319-kube-api-access-6vk5z\") on node \"crc\" DevicePath \"\"" Nov 24 18:29:42 crc kubenswrapper[4808]: I1124 18:29:42.242475 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b64ed14fe773da77e980a306ce0e524232b0a4d1f5816e610de5a0f9b4920638" Nov 24 18:29:42 crc kubenswrapper[4808]: I1124 18:29:42.242543 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/crc-debug-gxxrg" Nov 24 18:29:42 crc kubenswrapper[4808]: I1124 18:29:42.363471 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64482799-287e-4dea-b802-20ba3b468319" path="/var/lib/kubelet/pods/64482799-287e-4dea-b802-20ba3b468319/volumes" Nov 24 18:29:42 crc kubenswrapper[4808]: I1124 18:29:42.547796 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-d2lp6/crc-debug-gfk4h"] Nov 24 18:29:42 crc kubenswrapper[4808]: E1124 18:29:42.548336 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64482799-287e-4dea-b802-20ba3b468319" containerName="container-00" Nov 24 18:29:42 crc kubenswrapper[4808]: I1124 18:29:42.548361 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="64482799-287e-4dea-b802-20ba3b468319" containerName="container-00" Nov 24 18:29:42 crc kubenswrapper[4808]: I1124 18:29:42.548643 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="64482799-287e-4dea-b802-20ba3b468319" containerName="container-00" Nov 24 18:29:42 crc kubenswrapper[4808]: I1124 18:29:42.549464 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/crc-debug-gfk4h" Nov 24 18:29:42 crc kubenswrapper[4808]: I1124 18:29:42.688176 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxzxn\" (UniqueName: \"kubernetes.io/projected/e65bccbf-0da3-4855-941d-7123a2053275-kube-api-access-zxzxn\") pod \"crc-debug-gfk4h\" (UID: \"e65bccbf-0da3-4855-941d-7123a2053275\") " pod="openshift-must-gather-d2lp6/crc-debug-gfk4h" Nov 24 18:29:42 crc kubenswrapper[4808]: I1124 18:29:42.688296 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e65bccbf-0da3-4855-941d-7123a2053275-host\") pod \"crc-debug-gfk4h\" (UID: \"e65bccbf-0da3-4855-941d-7123a2053275\") " pod="openshift-must-gather-d2lp6/crc-debug-gfk4h" Nov 24 18:29:42 crc kubenswrapper[4808]: I1124 18:29:42.790400 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e65bccbf-0da3-4855-941d-7123a2053275-host\") pod \"crc-debug-gfk4h\" (UID: \"e65bccbf-0da3-4855-941d-7123a2053275\") " pod="openshift-must-gather-d2lp6/crc-debug-gfk4h" Nov 24 18:29:42 crc kubenswrapper[4808]: I1124 18:29:42.790639 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e65bccbf-0da3-4855-941d-7123a2053275-host\") pod \"crc-debug-gfk4h\" (UID: \"e65bccbf-0da3-4855-941d-7123a2053275\") " pod="openshift-must-gather-d2lp6/crc-debug-gfk4h" Nov 24 18:29:42 crc kubenswrapper[4808]: I1124 18:29:42.790682 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxzxn\" (UniqueName: \"kubernetes.io/projected/e65bccbf-0da3-4855-941d-7123a2053275-kube-api-access-zxzxn\") pod \"crc-debug-gfk4h\" (UID: \"e65bccbf-0da3-4855-941d-7123a2053275\") " pod="openshift-must-gather-d2lp6/crc-debug-gfk4h" Nov 24 18:29:42 crc kubenswrapper[4808]: I1124 18:29:42.868248 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxzxn\" (UniqueName: \"kubernetes.io/projected/e65bccbf-0da3-4855-941d-7123a2053275-kube-api-access-zxzxn\") pod \"crc-debug-gfk4h\" (UID: \"e65bccbf-0da3-4855-941d-7123a2053275\") " pod="openshift-must-gather-d2lp6/crc-debug-gfk4h" Nov 24 18:29:42 crc kubenswrapper[4808]: I1124 18:29:42.879745 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/crc-debug-gfk4h" Nov 24 18:29:43 crc kubenswrapper[4808]: I1124 18:29:43.252240 4808 generic.go:334] "Generic (PLEG): container finished" podID="e65bccbf-0da3-4855-941d-7123a2053275" containerID="a7d6cd2d44e11d5571387913c044699879aed47ea93c1745bd8e675a384715f7" exitCode=0 Nov 24 18:29:43 crc kubenswrapper[4808]: I1124 18:29:43.252332 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2lp6/crc-debug-gfk4h" event={"ID":"e65bccbf-0da3-4855-941d-7123a2053275","Type":"ContainerDied","Data":"a7d6cd2d44e11d5571387913c044699879aed47ea93c1745bd8e675a384715f7"} Nov 24 18:29:43 crc kubenswrapper[4808]: I1124 18:29:43.252515 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2lp6/crc-debug-gfk4h" event={"ID":"e65bccbf-0da3-4855-941d-7123a2053275","Type":"ContainerStarted","Data":"45e38eb9b75a4ac97a8154c40ad65119ad1a4f86785eac30b2712093395bea6f"} Nov 24 18:29:43 crc kubenswrapper[4808]: I1124 18:29:43.657976 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-d2lp6/crc-debug-gfk4h"] Nov 24 18:29:43 crc kubenswrapper[4808]: I1124 18:29:43.668853 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-d2lp6/crc-debug-gfk4h"] Nov 24 18:29:44 crc kubenswrapper[4808]: I1124 18:29:44.347297 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:29:44 crc kubenswrapper[4808]: E1124 18:29:44.347840 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:29:44 crc kubenswrapper[4808]: I1124 18:29:44.885470 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/crc-debug-gfk4h" Nov 24 18:29:44 crc kubenswrapper[4808]: I1124 18:29:44.914430 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-d2lp6/crc-debug-tvzp5"] Nov 24 18:29:44 crc kubenswrapper[4808]: E1124 18:29:44.914887 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e65bccbf-0da3-4855-941d-7123a2053275" containerName="container-00" Nov 24 18:29:44 crc kubenswrapper[4808]: I1124 18:29:44.914911 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e65bccbf-0da3-4855-941d-7123a2053275" containerName="container-00" Nov 24 18:29:44 crc kubenswrapper[4808]: I1124 18:29:44.915132 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e65bccbf-0da3-4855-941d-7123a2053275" containerName="container-00" Nov 24 18:29:44 crc kubenswrapper[4808]: I1124 18:29:44.915788 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/crc-debug-tvzp5" Nov 24 18:29:45 crc kubenswrapper[4808]: I1124 18:29:45.026135 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxzxn\" (UniqueName: \"kubernetes.io/projected/e65bccbf-0da3-4855-941d-7123a2053275-kube-api-access-zxzxn\") pod \"e65bccbf-0da3-4855-941d-7123a2053275\" (UID: \"e65bccbf-0da3-4855-941d-7123a2053275\") " Nov 24 18:29:45 crc kubenswrapper[4808]: I1124 18:29:45.026203 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e65bccbf-0da3-4855-941d-7123a2053275-host\") pod \"e65bccbf-0da3-4855-941d-7123a2053275\" (UID: \"e65bccbf-0da3-4855-941d-7123a2053275\") " Nov 24 18:29:45 crc kubenswrapper[4808]: I1124 18:29:45.026394 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e65bccbf-0da3-4855-941d-7123a2053275-host" (OuterVolumeSpecName: "host") pod "e65bccbf-0da3-4855-941d-7123a2053275" (UID: "e65bccbf-0da3-4855-941d-7123a2053275"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 18:29:45 crc kubenswrapper[4808]: I1124 18:29:45.026529 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/be36316c-baa8-4264-9483-97740876a0b8-host\") pod \"crc-debug-tvzp5\" (UID: \"be36316c-baa8-4264-9483-97740876a0b8\") " pod="openshift-must-gather-d2lp6/crc-debug-tvzp5" Nov 24 18:29:45 crc kubenswrapper[4808]: I1124 18:29:45.026613 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwgl8\" (UniqueName: \"kubernetes.io/projected/be36316c-baa8-4264-9483-97740876a0b8-kube-api-access-lwgl8\") pod \"crc-debug-tvzp5\" (UID: \"be36316c-baa8-4264-9483-97740876a0b8\") " pod="openshift-must-gather-d2lp6/crc-debug-tvzp5" Nov 24 18:29:45 crc kubenswrapper[4808]: I1124 18:29:45.026750 4808 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e65bccbf-0da3-4855-941d-7123a2053275-host\") on node \"crc\" DevicePath \"\"" Nov 24 18:29:45 crc kubenswrapper[4808]: I1124 18:29:45.032002 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e65bccbf-0da3-4855-941d-7123a2053275-kube-api-access-zxzxn" (OuterVolumeSpecName: "kube-api-access-zxzxn") pod "e65bccbf-0da3-4855-941d-7123a2053275" (UID: "e65bccbf-0da3-4855-941d-7123a2053275"). InnerVolumeSpecName "kube-api-access-zxzxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:29:45 crc kubenswrapper[4808]: I1124 18:29:45.128633 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/be36316c-baa8-4264-9483-97740876a0b8-host\") pod \"crc-debug-tvzp5\" (UID: \"be36316c-baa8-4264-9483-97740876a0b8\") " pod="openshift-must-gather-d2lp6/crc-debug-tvzp5" Nov 24 18:29:45 crc kubenswrapper[4808]: I1124 18:29:45.128782 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwgl8\" (UniqueName: \"kubernetes.io/projected/be36316c-baa8-4264-9483-97740876a0b8-kube-api-access-lwgl8\") pod \"crc-debug-tvzp5\" (UID: \"be36316c-baa8-4264-9483-97740876a0b8\") " pod="openshift-must-gather-d2lp6/crc-debug-tvzp5" Nov 24 18:29:45 crc kubenswrapper[4808]: I1124 18:29:45.128796 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/be36316c-baa8-4264-9483-97740876a0b8-host\") pod \"crc-debug-tvzp5\" (UID: \"be36316c-baa8-4264-9483-97740876a0b8\") " pod="openshift-must-gather-d2lp6/crc-debug-tvzp5" Nov 24 18:29:45 crc kubenswrapper[4808]: I1124 18:29:45.128925 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxzxn\" (UniqueName: \"kubernetes.io/projected/e65bccbf-0da3-4855-941d-7123a2053275-kube-api-access-zxzxn\") on node \"crc\" DevicePath \"\"" Nov 24 18:29:45 crc kubenswrapper[4808]: I1124 18:29:45.150395 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwgl8\" (UniqueName: \"kubernetes.io/projected/be36316c-baa8-4264-9483-97740876a0b8-kube-api-access-lwgl8\") pod \"crc-debug-tvzp5\" (UID: \"be36316c-baa8-4264-9483-97740876a0b8\") " pod="openshift-must-gather-d2lp6/crc-debug-tvzp5" Nov 24 18:29:45 crc kubenswrapper[4808]: I1124 18:29:45.231420 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/crc-debug-tvzp5" Nov 24 18:29:45 crc kubenswrapper[4808]: I1124 18:29:45.276617 4808 scope.go:117] "RemoveContainer" containerID="a7d6cd2d44e11d5571387913c044699879aed47ea93c1745bd8e675a384715f7" Nov 24 18:29:45 crc kubenswrapper[4808]: I1124 18:29:45.276680 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/crc-debug-gfk4h" Nov 24 18:29:46 crc kubenswrapper[4808]: I1124 18:29:46.287611 4808 generic.go:334] "Generic (PLEG): container finished" podID="be36316c-baa8-4264-9483-97740876a0b8" containerID="fdeb79b12f03ba676260be78e5d9d3f83cfa4e8f3db9448df9ea5c260f7102f0" exitCode=0 Nov 24 18:29:46 crc kubenswrapper[4808]: I1124 18:29:46.287661 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2lp6/crc-debug-tvzp5" event={"ID":"be36316c-baa8-4264-9483-97740876a0b8","Type":"ContainerDied","Data":"fdeb79b12f03ba676260be78e5d9d3f83cfa4e8f3db9448df9ea5c260f7102f0"} Nov 24 18:29:46 crc kubenswrapper[4808]: I1124 18:29:46.287689 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2lp6/crc-debug-tvzp5" event={"ID":"be36316c-baa8-4264-9483-97740876a0b8","Type":"ContainerStarted","Data":"3009147aca64e521a63b2a2871a5a6988ef49366d0356b94d20e7800de105c6c"} Nov 24 18:29:46 crc kubenswrapper[4808]: I1124 18:29:46.325703 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-d2lp6/crc-debug-tvzp5"] Nov 24 18:29:46 crc kubenswrapper[4808]: I1124 18:29:46.335444 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-d2lp6/crc-debug-tvzp5"] Nov 24 18:29:46 crc kubenswrapper[4808]: I1124 18:29:46.356993 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e65bccbf-0da3-4855-941d-7123a2053275" path="/var/lib/kubelet/pods/e65bccbf-0da3-4855-941d-7123a2053275/volumes" Nov 24 18:29:47 crc kubenswrapper[4808]: I1124 18:29:47.377513 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/crc-debug-tvzp5" Nov 24 18:29:47 crc kubenswrapper[4808]: I1124 18:29:47.471497 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwgl8\" (UniqueName: \"kubernetes.io/projected/be36316c-baa8-4264-9483-97740876a0b8-kube-api-access-lwgl8\") pod \"be36316c-baa8-4264-9483-97740876a0b8\" (UID: \"be36316c-baa8-4264-9483-97740876a0b8\") " Nov 24 18:29:47 crc kubenswrapper[4808]: I1124 18:29:47.471670 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/be36316c-baa8-4264-9483-97740876a0b8-host\") pod \"be36316c-baa8-4264-9483-97740876a0b8\" (UID: \"be36316c-baa8-4264-9483-97740876a0b8\") " Nov 24 18:29:47 crc kubenswrapper[4808]: I1124 18:29:47.471772 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be36316c-baa8-4264-9483-97740876a0b8-host" (OuterVolumeSpecName: "host") pod "be36316c-baa8-4264-9483-97740876a0b8" (UID: "be36316c-baa8-4264-9483-97740876a0b8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 18:29:47 crc kubenswrapper[4808]: I1124 18:29:47.472349 4808 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/be36316c-baa8-4264-9483-97740876a0b8-host\") on node \"crc\" DevicePath \"\"" Nov 24 18:29:47 crc kubenswrapper[4808]: I1124 18:29:47.479858 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be36316c-baa8-4264-9483-97740876a0b8-kube-api-access-lwgl8" (OuterVolumeSpecName: "kube-api-access-lwgl8") pod "be36316c-baa8-4264-9483-97740876a0b8" (UID: "be36316c-baa8-4264-9483-97740876a0b8"). InnerVolumeSpecName "kube-api-access-lwgl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:29:47 crc kubenswrapper[4808]: I1124 18:29:47.575008 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwgl8\" (UniqueName: \"kubernetes.io/projected/be36316c-baa8-4264-9483-97740876a0b8-kube-api-access-lwgl8\") on node \"crc\" DevicePath \"\"" Nov 24 18:29:48 crc kubenswrapper[4808]: I1124 18:29:48.307835 4808 scope.go:117] "RemoveContainer" containerID="fdeb79b12f03ba676260be78e5d9d3f83cfa4e8f3db9448df9ea5c260f7102f0" Nov 24 18:29:48 crc kubenswrapper[4808]: I1124 18:29:48.307888 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/crc-debug-tvzp5" Nov 24 18:29:48 crc kubenswrapper[4808]: I1124 18:29:48.375255 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be36316c-baa8-4264-9483-97740876a0b8" path="/var/lib/kubelet/pods/be36316c-baa8-4264-9483-97740876a0b8/volumes" Nov 24 18:29:55 crc kubenswrapper[4808]: I1124 18:29:55.347388 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:29:55 crc kubenswrapper[4808]: E1124 18:29:55.348140 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.197368 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5"] Nov 24 18:30:00 crc kubenswrapper[4808]: E1124 18:30:00.199605 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be36316c-baa8-4264-9483-97740876a0b8" containerName="container-00" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.199635 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="be36316c-baa8-4264-9483-97740876a0b8" containerName="container-00" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.203080 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="be36316c-baa8-4264-9483-97740876a0b8" containerName="container-00" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.205591 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.210785 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.211135 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.220688 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5"] Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.399188 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-secret-volume\") pod \"collect-profiles-29400150-f2ct5\" (UID: \"e0244192-db95-42b9-9dd0-dd4d31f5d5e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.399279 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-config-volume\") pod \"collect-profiles-29400150-f2ct5\" (UID: \"e0244192-db95-42b9-9dd0-dd4d31f5d5e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.399490 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4trht\" (UniqueName: \"kubernetes.io/projected/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-kube-api-access-4trht\") pod \"collect-profiles-29400150-f2ct5\" (UID: \"e0244192-db95-42b9-9dd0-dd4d31f5d5e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.502134 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4trht\" (UniqueName: \"kubernetes.io/projected/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-kube-api-access-4trht\") pod \"collect-profiles-29400150-f2ct5\" (UID: \"e0244192-db95-42b9-9dd0-dd4d31f5d5e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.503367 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-secret-volume\") pod \"collect-profiles-29400150-f2ct5\" (UID: \"e0244192-db95-42b9-9dd0-dd4d31f5d5e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.503468 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-config-volume\") pod \"collect-profiles-29400150-f2ct5\" (UID: \"e0244192-db95-42b9-9dd0-dd4d31f5d5e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.504865 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-config-volume\") pod \"collect-profiles-29400150-f2ct5\" (UID: \"e0244192-db95-42b9-9dd0-dd4d31f5d5e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.512619 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-secret-volume\") pod \"collect-profiles-29400150-f2ct5\" (UID: \"e0244192-db95-42b9-9dd0-dd4d31f5d5e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.525079 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4trht\" (UniqueName: \"kubernetes.io/projected/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-kube-api-access-4trht\") pod \"collect-profiles-29400150-f2ct5\" (UID: \"e0244192-db95-42b9-9dd0-dd4d31f5d5e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" Nov 24 18:30:00 crc kubenswrapper[4808]: I1124 18:30:00.548561 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" Nov 24 18:30:01 crc kubenswrapper[4808]: I1124 18:30:01.025743 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5"] Nov 24 18:30:01 crc kubenswrapper[4808]: I1124 18:30:01.440870 4808 generic.go:334] "Generic (PLEG): container finished" podID="e0244192-db95-42b9-9dd0-dd4d31f5d5e3" containerID="1135a5677748f557bcebd1f07db503ee6b3ad3577b2b5677570f74b75ddaff19" exitCode=0 Nov 24 18:30:01 crc kubenswrapper[4808]: I1124 18:30:01.440910 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" event={"ID":"e0244192-db95-42b9-9dd0-dd4d31f5d5e3","Type":"ContainerDied","Data":"1135a5677748f557bcebd1f07db503ee6b3ad3577b2b5677570f74b75ddaff19"} Nov 24 18:30:01 crc kubenswrapper[4808]: I1124 18:30:01.440935 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" event={"ID":"e0244192-db95-42b9-9dd0-dd4d31f5d5e3","Type":"ContainerStarted","Data":"4a8f829d5402e61f2bf61fb385fbeb2fb28a478d7e54a8fd6f58665da2e02ccf"} Nov 24 18:30:02 crc kubenswrapper[4808]: I1124 18:30:02.773910 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" Nov 24 18:30:02 crc kubenswrapper[4808]: I1124 18:30:02.949979 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4trht\" (UniqueName: \"kubernetes.io/projected/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-kube-api-access-4trht\") pod \"e0244192-db95-42b9-9dd0-dd4d31f5d5e3\" (UID: \"e0244192-db95-42b9-9dd0-dd4d31f5d5e3\") " Nov 24 18:30:02 crc kubenswrapper[4808]: I1124 18:30:02.950191 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-config-volume\") pod \"e0244192-db95-42b9-9dd0-dd4d31f5d5e3\" (UID: \"e0244192-db95-42b9-9dd0-dd4d31f5d5e3\") " Nov 24 18:30:02 crc kubenswrapper[4808]: I1124 18:30:02.950245 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-secret-volume\") pod \"e0244192-db95-42b9-9dd0-dd4d31f5d5e3\" (UID: \"e0244192-db95-42b9-9dd0-dd4d31f5d5e3\") " Nov 24 18:30:02 crc kubenswrapper[4808]: I1124 18:30:02.951037 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-config-volume" (OuterVolumeSpecName: "config-volume") pod "e0244192-db95-42b9-9dd0-dd4d31f5d5e3" (UID: "e0244192-db95-42b9-9dd0-dd4d31f5d5e3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 18:30:02 crc kubenswrapper[4808]: I1124 18:30:02.957905 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-kube-api-access-4trht" (OuterVolumeSpecName: "kube-api-access-4trht") pod "e0244192-db95-42b9-9dd0-dd4d31f5d5e3" (UID: "e0244192-db95-42b9-9dd0-dd4d31f5d5e3"). InnerVolumeSpecName "kube-api-access-4trht". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:30:02 crc kubenswrapper[4808]: I1124 18:30:02.958097 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e0244192-db95-42b9-9dd0-dd4d31f5d5e3" (UID: "e0244192-db95-42b9-9dd0-dd4d31f5d5e3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:30:03 crc kubenswrapper[4808]: I1124 18:30:03.051792 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4trht\" (UniqueName: \"kubernetes.io/projected/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-kube-api-access-4trht\") on node \"crc\" DevicePath \"\"" Nov 24 18:30:03 crc kubenswrapper[4808]: I1124 18:30:03.051830 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 18:30:03 crc kubenswrapper[4808]: I1124 18:30:03.051839 4808 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0244192-db95-42b9-9dd0-dd4d31f5d5e3-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 18:30:03 crc kubenswrapper[4808]: I1124 18:30:03.457786 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" event={"ID":"e0244192-db95-42b9-9dd0-dd4d31f5d5e3","Type":"ContainerDied","Data":"4a8f829d5402e61f2bf61fb385fbeb2fb28a478d7e54a8fd6f58665da2e02ccf"} Nov 24 18:30:03 crc kubenswrapper[4808]: I1124 18:30:03.457824 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a8f829d5402e61f2bf61fb385fbeb2fb28a478d7e54a8fd6f58665da2e02ccf" Nov 24 18:30:03 crc kubenswrapper[4808]: I1124 18:30:03.457878 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400150-f2ct5" Nov 24 18:30:03 crc kubenswrapper[4808]: I1124 18:30:03.843397 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp"] Nov 24 18:30:03 crc kubenswrapper[4808]: I1124 18:30:03.849233 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400105-xfvbp"] Nov 24 18:30:04 crc kubenswrapper[4808]: I1124 18:30:04.358751 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71bc118d-05fa-4b8e-84ff-e8b9ad75de68" path="/var/lib/kubelet/pods/71bc118d-05fa-4b8e-84ff-e8b9ad75de68/volumes" Nov 24 18:30:06 crc kubenswrapper[4808]: I1124 18:30:06.348355 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:30:06 crc kubenswrapper[4808]: E1124 18:30:06.349241 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:30:16 crc kubenswrapper[4808]: I1124 18:30:16.203670 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67f98cdffd-vs7fq_45f1a3c5-b0ad-429f-a2ff-5d19a23ca362/barbican-api/0.log" Nov 24 18:30:16 crc kubenswrapper[4808]: I1124 18:30:16.408935 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67f98cdffd-vs7fq_45f1a3c5-b0ad-429f-a2ff-5d19a23ca362/barbican-api-log/0.log" Nov 24 18:30:16 crc kubenswrapper[4808]: I1124 18:30:16.432212 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7549f65886-b8qq5_add8319e-0677-426b-b094-25b79b7a77e4/barbican-keystone-listener/0.log" Nov 24 18:30:16 crc kubenswrapper[4808]: I1124 18:30:16.487108 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7549f65886-b8qq5_add8319e-0677-426b-b094-25b79b7a77e4/barbican-keystone-listener-log/0.log" Nov 24 18:30:16 crc kubenswrapper[4808]: I1124 18:30:16.630215 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5db8598f6c-zs7cb_c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba/barbican-worker/0.log" Nov 24 18:30:16 crc kubenswrapper[4808]: I1124 18:30:16.649346 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5db8598f6c-zs7cb_c2a03c3d-9e8e-4713-97e9-e53e79e9e1ba/barbican-worker-log/0.log" Nov 24 18:30:16 crc kubenswrapper[4808]: I1124 18:30:16.833445 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-mxkzr_b6ec6116-79f6-4178-9ebc-bc29c07889b3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:16 crc kubenswrapper[4808]: I1124 18:30:16.874948 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a9b7ee10-3d3d-4755-bf68-0d745a8e4c67/ceilometer-central-agent/0.log" Nov 24 18:30:17 crc kubenswrapper[4808]: I1124 18:30:17.073890 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a9b7ee10-3d3d-4755-bf68-0d745a8e4c67/ceilometer-notification-agent/0.log" Nov 24 18:30:17 crc kubenswrapper[4808]: I1124 18:30:17.158226 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a9b7ee10-3d3d-4755-bf68-0d745a8e4c67/proxy-httpd/0.log" Nov 24 18:30:17 crc kubenswrapper[4808]: I1124 18:30:17.220996 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a9b7ee10-3d3d-4755-bf68-0d745a8e4c67/sg-core/0.log" Nov 24 18:30:17 crc kubenswrapper[4808]: I1124 18:30:17.309716 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_323bb785-3e02-469e-b169-22109303439e/cinder-api/0.log" Nov 24 18:30:17 crc kubenswrapper[4808]: I1124 18:30:17.368515 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_323bb785-3e02-469e-b169-22109303439e/cinder-api-log/0.log" Nov 24 18:30:17 crc kubenswrapper[4808]: I1124 18:30:17.445049 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a/cinder-scheduler/0.log" Nov 24 18:30:17 crc kubenswrapper[4808]: I1124 18:30:17.521946 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7eaf9bc6-bdea-4d20-a157-b7ef6ce2250a/probe/0.log" Nov 24 18:30:17 crc kubenswrapper[4808]: I1124 18:30:17.639756 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wglmk_91323238-4f1e-4dd3-84f2-6fc8bf252642/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:17 crc kubenswrapper[4808]: I1124 18:30:17.782276 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-h5fwk_faadd5b0-6e65-467e-b510-090ee77362b4/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:17 crc kubenswrapper[4808]: I1124 18:30:17.836814 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-m2sjz_db145eef-fe79-49c1-9fd1-570b506b3367/init/0.log" Nov 24 18:30:17 crc kubenswrapper[4808]: I1124 18:30:17.974586 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-m2sjz_db145eef-fe79-49c1-9fd1-570b506b3367/init/0.log" Nov 24 18:30:18 crc kubenswrapper[4808]: I1124 18:30:18.045535 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-m2sjz_db145eef-fe79-49c1-9fd1-570b506b3367/dnsmasq-dns/0.log" Nov 24 18:30:18 crc kubenswrapper[4808]: I1124 18:30:18.048732 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-s4f85_ee20a194-4626-4fe3-8138-e103a44a2006/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:18 crc kubenswrapper[4808]: I1124 18:30:18.222980 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_dc25c6d7-a302-4ea0-bd25-2b19f9b17a71/glance-log/0.log" Nov 24 18:30:18 crc kubenswrapper[4808]: I1124 18:30:18.236168 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_dc25c6d7-a302-4ea0-bd25-2b19f9b17a71/glance-httpd/0.log" Nov 24 18:30:18 crc kubenswrapper[4808]: I1124 18:30:18.345963 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_88cb8610-eb5e-45de-8028-37fc361beaac/glance-httpd/0.log" Nov 24 18:30:18 crc kubenswrapper[4808]: I1124 18:30:18.353333 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:30:18 crc kubenswrapper[4808]: E1124 18:30:18.353590 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:30:18 crc kubenswrapper[4808]: I1124 18:30:18.420196 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_88cb8610-eb5e-45de-8028-37fc361beaac/glance-log/0.log" Nov 24 18:30:18 crc kubenswrapper[4808]: I1124 18:30:18.714711 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-f8xrf_91e767d6-f4dd-4bf8-b58d-37dc9073e90f/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:18 crc kubenswrapper[4808]: I1124 18:30:18.749189 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5ddbdb6df8-pv5zj_96142f0a-3231-4617-bbd2-9f440c73908e/horizon/0.log" Nov 24 18:30:18 crc kubenswrapper[4808]: I1124 18:30:18.910709 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-f8mg9_37c9ab78-83a5-4432-8c14-b1e973891168/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:18 crc kubenswrapper[4808]: I1124 18:30:18.984975 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5ddbdb6df8-pv5zj_96142f0a-3231-4617-bbd2-9f440c73908e/horizon-log/0.log" Nov 24 18:30:19 crc kubenswrapper[4808]: I1124 18:30:19.136745 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29400121-9z8sf_f7d529f4-bce4-4d3a-83b2-38c65e0f4e7e/keystone-cron/0.log" Nov 24 18:30:19 crc kubenswrapper[4808]: I1124 18:30:19.197216 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7b7d8889-c8cpc_3c94ac11-4a54-4fde-b2dd-e9e92d58894f/keystone-api/0.log" Nov 24 18:30:19 crc kubenswrapper[4808]: I1124 18:30:19.326677 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_0c8560e5-eccf-482d-bc1b-74e44ccfeff1/kube-state-metrics/0.log" Nov 24 18:30:19 crc kubenswrapper[4808]: I1124 18:30:19.363702 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-7gcd2_3a26fcdc-f080-4677-a308-b08584153734/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:19 crc kubenswrapper[4808]: I1124 18:30:19.794997 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-ff547b7f9-dfqk9_76a5234c-3951-4c26-92c2-0ead15585dc5/neutron-httpd/0.log" Nov 24 18:30:19 crc kubenswrapper[4808]: I1124 18:30:19.797411 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-ff547b7f9-dfqk9_76a5234c-3951-4c26-92c2-0ead15585dc5/neutron-api/0.log" Nov 24 18:30:19 crc kubenswrapper[4808]: I1124 18:30:19.977213 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-bpnvh_40e7c1c1-2df4-4516-a0b7-ec3ab8ed723f/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:20 crc kubenswrapper[4808]: I1124 18:30:20.467832 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2f6861ff-4148-4476-8a97-ef9228b91c8b/nova-api-log/0.log" Nov 24 18:30:20 crc kubenswrapper[4808]: I1124 18:30:20.530705 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_1c4e90b9-4ce7-4931-922f-ce473cc9d5ed/nova-cell0-conductor-conductor/0.log" Nov 24 18:30:20 crc kubenswrapper[4808]: I1124 18:30:20.874712 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2f6861ff-4148-4476-8a97-ef9228b91c8b/nova-api-api/0.log" Nov 24 18:30:20 crc kubenswrapper[4808]: I1124 18:30:20.882466 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_7074b2db-ae2b-4fee-a890-6cc64cd46cf8/nova-cell1-conductor-conductor/0.log" Nov 24 18:30:21 crc kubenswrapper[4808]: I1124 18:30:21.060651 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_d32aec99-9b13-4e2e-af1a-9b55ab86026b/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 18:30:21 crc kubenswrapper[4808]: I1124 18:30:21.183397 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-lfxmh_e03cdbd6-8f86-468f-a798-961c6a4920d7/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:21 crc kubenswrapper[4808]: I1124 18:30:21.345166 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e12cb986-dbb8-42c6-a6f4-7d49d49107d6/nova-metadata-log/0.log" Nov 24 18:30:21 crc kubenswrapper[4808]: I1124 18:30:21.572197 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1631f202-c106-4965-874f-7af292288025/nova-scheduler-scheduler/0.log" Nov 24 18:30:21 crc kubenswrapper[4808]: I1124 18:30:21.607644 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_73a668e5-4465-455b-a109-3fcee8ea831f/mysql-bootstrap/0.log" Nov 24 18:30:21 crc kubenswrapper[4808]: I1124 18:30:21.838984 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_73a668e5-4465-455b-a109-3fcee8ea831f/mysql-bootstrap/0.log" Nov 24 18:30:21 crc kubenswrapper[4808]: I1124 18:30:21.853951 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_73a668e5-4465-455b-a109-3fcee8ea831f/galera/0.log" Nov 24 18:30:22 crc kubenswrapper[4808]: I1124 18:30:22.076288 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ea9156df-b46b-43f4-97b8-6f1aaf74b6f6/mysql-bootstrap/0.log" Nov 24 18:30:22 crc kubenswrapper[4808]: I1124 18:30:22.338268 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ea9156df-b46b-43f4-97b8-6f1aaf74b6f6/mysql-bootstrap/0.log" Nov 24 18:30:22 crc kubenswrapper[4808]: I1124 18:30:22.345839 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ea9156df-b46b-43f4-97b8-6f1aaf74b6f6/galera/0.log" Nov 24 18:30:22 crc kubenswrapper[4808]: I1124 18:30:22.510828 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2fb472ac-c08a-4ed0-b1eb-ab83eaaa8c6c/openstackclient/0.log" Nov 24 18:30:22 crc kubenswrapper[4808]: I1124 18:30:22.551699 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e12cb986-dbb8-42c6-a6f4-7d49d49107d6/nova-metadata-metadata/0.log" Nov 24 18:30:22 crc kubenswrapper[4808]: I1124 18:30:22.593348 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kp2db_a3ac52cd-4f6a-4360-867b-8d0a156089b4/openstack-network-exporter/0.log" Nov 24 18:30:22 crc kubenswrapper[4808]: I1124 18:30:22.753720 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-58sfc_e091dc40-a06a-482a-873c-7be28cd7fd29/ovsdb-server-init/0.log" Nov 24 18:30:22 crc kubenswrapper[4808]: I1124 18:30:22.923353 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-58sfc_e091dc40-a06a-482a-873c-7be28cd7fd29/ovsdb-server-init/0.log" Nov 24 18:30:22 crc kubenswrapper[4808]: I1124 18:30:22.949518 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-58sfc_e091dc40-a06a-482a-873c-7be28cd7fd29/ovs-vswitchd/0.log" Nov 24 18:30:23 crc kubenswrapper[4808]: I1124 18:30:23.003355 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-58sfc_e091dc40-a06a-482a-873c-7be28cd7fd29/ovsdb-server/0.log" Nov 24 18:30:23 crc kubenswrapper[4808]: I1124 18:30:23.132783 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-zrsgl_dc7ca1ba-7fd1-40aa-8eda-8ee9a69f10ec/ovn-controller/0.log" Nov 24 18:30:23 crc kubenswrapper[4808]: I1124 18:30:23.268722 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-v5zl8_119c4dcd-dc15-44aa-bfde-1a2931d8b83d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:23 crc kubenswrapper[4808]: I1124 18:30:23.333204 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a8ca4d50-8de1-4d3a-b5ed-313d024d224a/openstack-network-exporter/0.log" Nov 24 18:30:23 crc kubenswrapper[4808]: I1124 18:30:23.357628 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a8ca4d50-8de1-4d3a-b5ed-313d024d224a/ovn-northd/0.log" Nov 24 18:30:23 crc kubenswrapper[4808]: I1124 18:30:23.526981 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_dc135c1e-57ee-4be6-835e-b9341e7226e8/openstack-network-exporter/0.log" Nov 24 18:30:23 crc kubenswrapper[4808]: I1124 18:30:23.552952 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_dc135c1e-57ee-4be6-835e-b9341e7226e8/ovsdbserver-nb/0.log" Nov 24 18:30:23 crc kubenswrapper[4808]: I1124 18:30:23.701842 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_67b93485-6ca7-4192-901b-581dbf985068/openstack-network-exporter/0.log" Nov 24 18:30:23 crc kubenswrapper[4808]: I1124 18:30:23.775977 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_67b93485-6ca7-4192-901b-581dbf985068/ovsdbserver-sb/0.log" Nov 24 18:30:23 crc kubenswrapper[4808]: I1124 18:30:23.819593 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-785cbcd98d-bqfnp_18412693-fc60-4860-a2b2-75e830b495a7/placement-api/0.log" Nov 24 18:30:24 crc kubenswrapper[4808]: I1124 18:30:24.138682 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-785cbcd98d-bqfnp_18412693-fc60-4860-a2b2-75e830b495a7/placement-log/0.log" Nov 24 18:30:24 crc kubenswrapper[4808]: I1124 18:30:24.141944 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_78791f0f-ea96-45ba-9735-9d10ddebea10/setup-container/0.log" Nov 24 18:30:24 crc kubenswrapper[4808]: I1124 18:30:24.382775 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_78791f0f-ea96-45ba-9735-9d10ddebea10/rabbitmq/0.log" Nov 24 18:30:24 crc kubenswrapper[4808]: I1124 18:30:24.407990 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bddeab55-cf79-4073-b5c4-420508603e10/setup-container/0.log" Nov 24 18:30:24 crc kubenswrapper[4808]: I1124 18:30:24.473817 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_78791f0f-ea96-45ba-9735-9d10ddebea10/setup-container/0.log" Nov 24 18:30:24 crc kubenswrapper[4808]: I1124 18:30:24.654834 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bddeab55-cf79-4073-b5c4-420508603e10/setup-container/0.log" Nov 24 18:30:24 crc kubenswrapper[4808]: I1124 18:30:24.743548 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bddeab55-cf79-4073-b5c4-420508603e10/rabbitmq/0.log" Nov 24 18:30:24 crc kubenswrapper[4808]: I1124 18:30:24.793717 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-zsk5s_2a9a918d-83b1-4e41-89a1-15e375a5f18e/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:24 crc kubenswrapper[4808]: I1124 18:30:24.953007 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-zkfw7_27ee00de-fd0b-435d-84a5-28cf3c4d0dc7/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:25 crc kubenswrapper[4808]: I1124 18:30:25.024112 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-wvqwl_a36913c9-4487-4403-9797-8ebacf1304d7/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:25 crc kubenswrapper[4808]: I1124 18:30:25.186935 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-qztqt_92d01f67-8a79-4e09-afad-bf04575f554b/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:25 crc kubenswrapper[4808]: I1124 18:30:25.244253 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-h2xqn_6bab5c2c-7363-40d3-855e-de98d478a482/ssh-known-hosts-edpm-deployment/0.log" Nov 24 18:30:25 crc kubenswrapper[4808]: I1124 18:30:25.478305 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-84b96bd867-hvzp6_a525134b-d25e-4a38-82ec-c7256c6d6697/proxy-server/0.log" Nov 24 18:30:25 crc kubenswrapper[4808]: I1124 18:30:25.545807 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-84b96bd867-hvzp6_a525134b-d25e-4a38-82ec-c7256c6d6697/proxy-httpd/0.log" Nov 24 18:30:25 crc kubenswrapper[4808]: I1124 18:30:25.600669 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-6ttz5_ee5479fc-5360-4541-990d-c8d1a97bae29/swift-ring-rebalance/0.log" Nov 24 18:30:25 crc kubenswrapper[4808]: I1124 18:30:25.728327 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/account-auditor/0.log" Nov 24 18:30:25 crc kubenswrapper[4808]: I1124 18:30:25.819490 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/account-reaper/0.log" Nov 24 18:30:25 crc kubenswrapper[4808]: I1124 18:30:25.877834 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/account-replicator/0.log" Nov 24 18:30:25 crc kubenswrapper[4808]: I1124 18:30:25.920001 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/account-server/0.log" Nov 24 18:30:25 crc kubenswrapper[4808]: I1124 18:30:25.946931 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/container-auditor/0.log" Nov 24 18:30:26 crc kubenswrapper[4808]: I1124 18:30:26.069147 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/container-replicator/0.log" Nov 24 18:30:26 crc kubenswrapper[4808]: I1124 18:30:26.093483 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/container-server/0.log" Nov 24 18:30:26 crc kubenswrapper[4808]: I1124 18:30:26.113814 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/container-updater/0.log" Nov 24 18:30:26 crc kubenswrapper[4808]: I1124 18:30:26.171834 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/object-auditor/0.log" Nov 24 18:30:26 crc kubenswrapper[4808]: I1124 18:30:26.283633 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/object-expirer/0.log" Nov 24 18:30:26 crc kubenswrapper[4808]: I1124 18:30:26.290869 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/object-replicator/0.log" Nov 24 18:30:26 crc kubenswrapper[4808]: I1124 18:30:26.321669 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/object-server/0.log" Nov 24 18:30:26 crc kubenswrapper[4808]: I1124 18:30:26.398728 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/object-updater/0.log" Nov 24 18:30:26 crc kubenswrapper[4808]: I1124 18:30:26.450411 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/rsync/0.log" Nov 24 18:30:26 crc kubenswrapper[4808]: I1124 18:30:26.514083 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c6e7dc0a-09ef-49f9-ae1d-b4bff6dfbb14/swift-recon-cron/0.log" Nov 24 18:30:26 crc kubenswrapper[4808]: I1124 18:30:26.631665 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-kt77n_7f556606-72e2-4fc7-bb57-cc4bd96c4f9b/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:26 crc kubenswrapper[4808]: I1124 18:30:26.794009 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_d014b1af-f795-43e4-af00-0a152d0cb945/tempest-tests-tempest-tests-runner/0.log" Nov 24 18:30:26 crc kubenswrapper[4808]: I1124 18:30:26.845771 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_5c95e77e-1080-4ca8-96d8-8b34632303ea/test-operator-logs-container/0.log" Nov 24 18:30:26 crc kubenswrapper[4808]: I1124 18:30:26.955741 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-z92sj_579d95f1-e760-4476-bc98-6e33dabf6780/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:30:30 crc kubenswrapper[4808]: I1124 18:30:30.347862 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:30:30 crc kubenswrapper[4808]: E1124 18:30:30.348376 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:30:35 crc kubenswrapper[4808]: I1124 18:30:35.135192 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_3bd73c68-0bbb-4026-b240-637e5b7c17b5/memcached/0.log" Nov 24 18:30:41 crc kubenswrapper[4808]: I1124 18:30:41.347761 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:30:41 crc kubenswrapper[4808]: E1124 18:30:41.349203 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:30:50 crc kubenswrapper[4808]: I1124 18:30:50.967578 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-bbdb9b8fd-hfwr9_2ece0883-b2fd-4309-b4bd-7639d95445e2/kube-rbac-proxy/0.log" Nov 24 18:30:51 crc kubenswrapper[4808]: I1124 18:30:51.015222 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-bbdb9b8fd-hfwr9_2ece0883-b2fd-4309-b4bd-7639d95445e2/manager/0.log" Nov 24 18:30:51 crc kubenswrapper[4808]: I1124 18:30:51.159903 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-942s7_bf340213-3bb6-4fc3-b8ed-9728393e9a5a/kube-rbac-proxy/0.log" Nov 24 18:30:51 crc kubenswrapper[4808]: I1124 18:30:51.239685 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-942s7_bf340213-3bb6-4fc3-b8ed-9728393e9a5a/manager/0.log" Nov 24 18:30:51 crc kubenswrapper[4808]: I1124 18:30:51.330204 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-nv6j9_edd12c74-ba4f-4cb4-8531-b98607fdab41/kube-rbac-proxy/0.log" Nov 24 18:30:51 crc kubenswrapper[4808]: I1124 18:30:51.368109 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-nv6j9_edd12c74-ba4f-4cb4-8531-b98607fdab41/manager/0.log" Nov 24 18:30:51 crc kubenswrapper[4808]: I1124 18:30:51.485038 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6_eb16e3fe-9175-449c-af90-76b3d2475753/util/0.log" Nov 24 18:30:51 crc kubenswrapper[4808]: I1124 18:30:51.653323 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6_eb16e3fe-9175-449c-af90-76b3d2475753/util/0.log" Nov 24 18:30:51 crc kubenswrapper[4808]: I1124 18:30:51.689941 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6_eb16e3fe-9175-449c-af90-76b3d2475753/pull/0.log" Nov 24 18:30:51 crc kubenswrapper[4808]: I1124 18:30:51.718420 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6_eb16e3fe-9175-449c-af90-76b3d2475753/pull/0.log" Nov 24 18:30:51 crc kubenswrapper[4808]: I1124 18:30:51.899309 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6_eb16e3fe-9175-449c-af90-76b3d2475753/pull/0.log" Nov 24 18:30:51 crc kubenswrapper[4808]: I1124 18:30:51.918113 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6_eb16e3fe-9175-449c-af90-76b3d2475753/util/0.log" Nov 24 18:30:51 crc kubenswrapper[4808]: I1124 18:30:51.918342 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f4c585506ee7b9544d8975ca8f41cd8090c1bfa1625d93eb3e6de8658dmkbs6_eb16e3fe-9175-449c-af90-76b3d2475753/extract/0.log" Nov 24 18:30:52 crc kubenswrapper[4808]: I1124 18:30:52.130087 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7fd7485997-ssvg7_fa1e211d-28e9-4ea1-beb5-6e3d0f5b8b9a/manager/0.log" Nov 24 18:30:52 crc kubenswrapper[4808]: I1124 18:30:52.152563 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7fd7485997-ssvg7_fa1e211d-28e9-4ea1-beb5-6e3d0f5b8b9a/kube-rbac-proxy/0.log" Nov 24 18:30:52 crc kubenswrapper[4808]: I1124 18:30:52.175224 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-758c497787-nnb8j_55559441-f0b0-4353-bbcb-55342cf6ef53/kube-rbac-proxy/0.log" Nov 24 18:30:52 crc kubenswrapper[4808]: I1124 18:30:52.363390 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-758c497787-nnb8j_55559441-f0b0-4353-bbcb-55342cf6ef53/manager/0.log" Nov 24 18:30:52 crc kubenswrapper[4808]: I1124 18:30:52.369505 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-2fzkr_352d8a11-a201-4bc3-8d10-8378a5e04bfb/kube-rbac-proxy/0.log" Nov 24 18:30:52 crc kubenswrapper[4808]: I1124 18:30:52.388788 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-2fzkr_352d8a11-a201-4bc3-8d10-8378a5e04bfb/manager/0.log" Nov 24 18:30:52 crc kubenswrapper[4808]: I1124 18:30:52.547873 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-299sb_58b1610a-0122-447e-a0a0-c2cb927a571f/kube-rbac-proxy/0.log" Nov 24 18:30:52 crc kubenswrapper[4808]: I1124 18:30:52.744292 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-q5xs5_58d593d8-712e-4eb5-8b1c-5df55db3f1d9/kube-rbac-proxy/0.log" Nov 24 18:30:52 crc kubenswrapper[4808]: I1124 18:30:52.750061 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-q5xs5_58d593d8-712e-4eb5-8b1c-5df55db3f1d9/manager/0.log" Nov 24 18:30:52 crc kubenswrapper[4808]: I1124 18:30:52.769016 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-299sb_58b1610a-0122-447e-a0a0-c2cb927a571f/manager/0.log" Nov 24 18:30:52 crc kubenswrapper[4808]: I1124 18:30:52.904749 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-tg4qk_9d1d2914-59b8-4660-80a4-d2a1cc17c2bc/kube-rbac-proxy/0.log" Nov 24 18:30:52 crc kubenswrapper[4808]: I1124 18:30:52.984040 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-tg4qk_9d1d2914-59b8-4660-80a4-d2a1cc17c2bc/manager/0.log" Nov 24 18:30:53 crc kubenswrapper[4808]: I1124 18:30:53.008101 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-8659db65fc-r4hcb_01054a01-64a1-4c81-b03f-5dfdf55ee4cf/kube-rbac-proxy/0.log" Nov 24 18:30:53 crc kubenswrapper[4808]: I1124 18:30:53.093699 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-8659db65fc-r4hcb_01054a01-64a1-4c81-b03f-5dfdf55ee4cf/manager/0.log" Nov 24 18:30:53 crc kubenswrapper[4808]: I1124 18:30:53.201836 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5f598f884b-xqgvh_ac1adbdb-b43a-4d5a-b599-7e2993f72a86/manager/0.log" Nov 24 18:30:53 crc kubenswrapper[4808]: I1124 18:30:53.202055 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5f598f884b-xqgvh_ac1adbdb-b43a-4d5a-b599-7e2993f72a86/kube-rbac-proxy/0.log" Nov 24 18:30:53 crc kubenswrapper[4808]: I1124 18:30:53.353029 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-qrr4f_8a4f09fc-1405-4a74-be87-1b94ee06742a/kube-rbac-proxy/0.log" Nov 24 18:30:53 crc kubenswrapper[4808]: I1124 18:30:53.383651 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-qrr4f_8a4f09fc-1405-4a74-be87-1b94ee06742a/manager/0.log" Nov 24 18:30:53 crc kubenswrapper[4808]: I1124 18:30:53.491006 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-rcwzz_47c249db-b332-44cf-baa5-811e04500c0f/kube-rbac-proxy/0.log" Nov 24 18:30:53 crc kubenswrapper[4808]: I1124 18:30:53.602252 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-b5vdf_af79a806-a6fb-40a1-8ac9-3fccd532c1fc/kube-rbac-proxy/0.log" Nov 24 18:30:53 crc kubenswrapper[4808]: I1124 18:30:53.602734 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-rcwzz_47c249db-b332-44cf-baa5-811e04500c0f/manager/0.log" Nov 24 18:30:53 crc kubenswrapper[4808]: I1124 18:30:53.663546 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-b5vdf_af79a806-a6fb-40a1-8ac9-3fccd532c1fc/manager/0.log" Nov 24 18:30:53 crc kubenswrapper[4808]: I1124 18:30:53.832753 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7d654b7b69f77rn_c796b113-a891-48b2-9a11-50a2577baf7e/kube-rbac-proxy/0.log" Nov 24 18:30:53 crc kubenswrapper[4808]: I1124 18:30:53.832994 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7d654b7b69f77rn_c796b113-a891-48b2-9a11-50a2577baf7e/manager/0.log" Nov 24 18:30:54 crc kubenswrapper[4808]: I1124 18:30:54.189061 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-798b7b4bc9-dhdfm_8d18f465-70ea-4d1f-8531-dcfee37c960d/operator/0.log" Nov 24 18:30:54 crc kubenswrapper[4808]: I1124 18:30:54.270202 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-d6xmw_24eb3123-6595-48d6-8d4a-f984d9a9e601/registry-server/0.log" Nov 24 18:30:54 crc kubenswrapper[4808]: I1124 18:30:54.366293 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:30:54 crc kubenswrapper[4808]: E1124 18:30:54.367451 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:30:54 crc kubenswrapper[4808]: I1124 18:30:54.444253 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-668bd84cc7-lm5c4_dc05be1a-a87d-4829-b2a8-c9d454b7d343/kube-rbac-proxy/0.log" Nov 24 18:30:54 crc kubenswrapper[4808]: I1124 18:30:54.541296 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-668bd84cc7-lm5c4_dc05be1a-a87d-4829-b2a8-c9d454b7d343/manager/0.log" Nov 24 18:30:54 crc kubenswrapper[4808]: I1124 18:30:54.618155 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-zgnpn_321df5bf-732b-4982-841b-3a9a0ddb73f2/kube-rbac-proxy/0.log" Nov 24 18:30:54 crc kubenswrapper[4808]: I1124 18:30:54.795615 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-zgnpn_321df5bf-732b-4982-841b-3a9a0ddb73f2/manager/0.log" Nov 24 18:30:54 crc kubenswrapper[4808]: I1124 18:30:54.876692 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-4r97r_5d6ac280-f3af-4ee8-8933-928ed0c501fc/operator/0.log" Nov 24 18:30:55 crc kubenswrapper[4808]: I1124 18:30:55.019636 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-mdswb_2c56d9f3-3ab5-4f3c-b250-e926807dbfc3/kube-rbac-proxy/0.log" Nov 24 18:30:55 crc kubenswrapper[4808]: I1124 18:30:55.073969 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-mdswb_2c56d9f3-3ab5-4f3c-b250-e926807dbfc3/manager/0.log" Nov 24 18:30:55 crc kubenswrapper[4808]: I1124 18:30:55.099771 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-9wwt6_ec05f4fb-ca34-4f68-b94e-95b08b21d57b/kube-rbac-proxy/0.log" Nov 24 18:30:55 crc kubenswrapper[4808]: I1124 18:30:55.217002 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-79c6ffd6b7-vvfcw_e597e7fc-3b56-493b-bcda-10669936636a/manager/0.log" Nov 24 18:30:55 crc kubenswrapper[4808]: I1124 18:30:55.318707 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-9wwt6_ec05f4fb-ca34-4f68-b94e-95b08b21d57b/manager/0.log" Nov 24 18:30:55 crc kubenswrapper[4808]: I1124 18:30:55.344634 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-66sz4_eb8cf6b6-33aa-427e-9d07-9f95c58c7d30/kube-rbac-proxy/0.log" Nov 24 18:30:55 crc kubenswrapper[4808]: I1124 18:30:55.373383 4808 scope.go:117] "RemoveContainer" containerID="a3737b255ab02d41b6d3f0f2d908df7352943c3f48369d2be026048a147c587b" Nov 24 18:30:55 crc kubenswrapper[4808]: I1124 18:30:55.406432 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-66sz4_eb8cf6b6-33aa-427e-9d07-9f95c58c7d30/manager/0.log" Nov 24 18:30:55 crc kubenswrapper[4808]: I1124 18:30:55.473345 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-wknfh_c62f31fd-04d1-48cc-ae7a-98e1e64c5af7/kube-rbac-proxy/0.log" Nov 24 18:30:55 crc kubenswrapper[4808]: I1124 18:30:55.514215 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-wknfh_c62f31fd-04d1-48cc-ae7a-98e1e64c5af7/manager/0.log" Nov 24 18:31:08 crc kubenswrapper[4808]: I1124 18:31:08.358806 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:31:08 crc kubenswrapper[4808]: E1124 18:31:08.359964 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:31:12 crc kubenswrapper[4808]: I1124 18:31:12.906079 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-ljn6r_4319fe26-e7cc-4665-881a-6275268b6330/control-plane-machine-set-operator/0.log" Nov 24 18:31:13 crc kubenswrapper[4808]: I1124 18:31:13.036975 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qrcr6_d54d24ca-c6c4-4313-8f22-908ce8939026/kube-rbac-proxy/0.log" Nov 24 18:31:13 crc kubenswrapper[4808]: I1124 18:31:13.069241 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qrcr6_d54d24ca-c6c4-4313-8f22-908ce8939026/machine-api-operator/0.log" Nov 24 18:31:20 crc kubenswrapper[4808]: I1124 18:31:20.347623 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:31:20 crc kubenswrapper[4808]: E1124 18:31:20.348533 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:31:26 crc kubenswrapper[4808]: I1124 18:31:26.994071 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-2qsx7_016a9f20-1203-4241-9210-55b50eeb60a1/cert-manager-controller/0.log" Nov 24 18:31:27 crc kubenswrapper[4808]: I1124 18:31:27.198250 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-cbqfz_24eadecc-5782-4e5b-8982-46d3e99fa5b1/cert-manager-cainjector/0.log" Nov 24 18:31:27 crc kubenswrapper[4808]: I1124 18:31:27.271208 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-bjlpj_a1ea24f0-1b28-4648-a09e-8c0539792ea1/cert-manager-webhook/0.log" Nov 24 18:31:31 crc kubenswrapper[4808]: I1124 18:31:31.347743 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:31:31 crc kubenswrapper[4808]: E1124 18:31:31.348653 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:31:41 crc kubenswrapper[4808]: I1124 18:31:41.954154 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-4x67c_b9930406-6bd1-4674-a697-e915c209fc08/nmstate-console-plugin/0.log" Nov 24 18:31:42 crc kubenswrapper[4808]: I1124 18:31:42.107611 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-mf5qh_125495c2-7b26-44c8-913a-0b98c6c81c98/kube-rbac-proxy/0.log" Nov 24 18:31:42 crc kubenswrapper[4808]: I1124 18:31:42.114622 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-kz6lk_9cc7b294-4f80-47be-88f4-e3b0c010ca1d/nmstate-handler/0.log" Nov 24 18:31:42 crc kubenswrapper[4808]: I1124 18:31:42.201026 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-mf5qh_125495c2-7b26-44c8-913a-0b98c6c81c98/nmstate-metrics/0.log" Nov 24 18:31:42 crc kubenswrapper[4808]: I1124 18:31:42.352497 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-6dp5q_48267952-ed56-4a20-98c9-e22f04e52c63/nmstate-operator/0.log" Nov 24 18:31:42 crc kubenswrapper[4808]: I1124 18:31:42.408116 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-rzhnm_b1fd2203-60a6-4172-bb4e-124fc185ade8/nmstate-webhook/0.log" Nov 24 18:31:46 crc kubenswrapper[4808]: I1124 18:31:46.348128 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:31:46 crc kubenswrapper[4808]: E1124 18:31:46.348872 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:31:58 crc kubenswrapper[4808]: I1124 18:31:58.079136 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-88twg_9afaca72-2a60-4b1f-a377-4510b24b887e/kube-rbac-proxy/0.log" Nov 24 18:31:58 crc kubenswrapper[4808]: I1124 18:31:58.177358 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-88twg_9afaca72-2a60-4b1f-a377-4510b24b887e/controller/0.log" Nov 24 18:31:58 crc kubenswrapper[4808]: I1124 18:31:58.310229 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-frr-files/0.log" Nov 24 18:31:58 crc kubenswrapper[4808]: I1124 18:31:58.475552 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-metrics/0.log" Nov 24 18:31:58 crc kubenswrapper[4808]: I1124 18:31:58.497241 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-frr-files/0.log" Nov 24 18:31:58 crc kubenswrapper[4808]: I1124 18:31:58.557973 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-reloader/0.log" Nov 24 18:31:58 crc kubenswrapper[4808]: I1124 18:31:58.573685 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-reloader/0.log" Nov 24 18:31:58 crc kubenswrapper[4808]: I1124 18:31:58.722999 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-frr-files/0.log" Nov 24 18:31:58 crc kubenswrapper[4808]: I1124 18:31:58.750513 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-metrics/0.log" Nov 24 18:31:58 crc kubenswrapper[4808]: I1124 18:31:58.750868 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-reloader/0.log" Nov 24 18:31:58 crc kubenswrapper[4808]: I1124 18:31:58.786397 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-metrics/0.log" Nov 24 18:31:58 crc kubenswrapper[4808]: I1124 18:31:58.896363 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-frr-files/0.log" Nov 24 18:31:58 crc kubenswrapper[4808]: I1124 18:31:58.922543 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-reloader/0.log" Nov 24 18:31:58 crc kubenswrapper[4808]: I1124 18:31:58.958808 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/cp-metrics/0.log" Nov 24 18:31:59 crc kubenswrapper[4808]: I1124 18:31:58.983217 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/controller/0.log" Nov 24 18:31:59 crc kubenswrapper[4808]: I1124 18:31:59.123750 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/frr-metrics/0.log" Nov 24 18:31:59 crc kubenswrapper[4808]: I1124 18:31:59.172364 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/kube-rbac-proxy/0.log" Nov 24 18:31:59 crc kubenswrapper[4808]: I1124 18:31:59.229452 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/kube-rbac-proxy-frr/0.log" Nov 24 18:31:59 crc kubenswrapper[4808]: I1124 18:31:59.395150 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/reloader/0.log" Nov 24 18:31:59 crc kubenswrapper[4808]: I1124 18:31:59.427258 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-9qp89_f5d3c911-b65c-46dc-a43a-60247b4d10f3/frr-k8s-webhook-server/0.log" Nov 24 18:31:59 crc kubenswrapper[4808]: I1124 18:31:59.628575 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-56c4766567-nztcn_8dc3fedf-199a-4446-a18a-3dd0f6818f8d/manager/0.log" Nov 24 18:31:59 crc kubenswrapper[4808]: I1124 18:31:59.801660 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7c6ddcbc67-pn9v5_da67a1e6-9fd1-4b27-b965-c907e51b6ce4/webhook-server/0.log" Nov 24 18:31:59 crc kubenswrapper[4808]: I1124 18:31:59.909900 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6mvpd_6f3367d5-54fd-4556-b1ba-5311e15433f3/kube-rbac-proxy/0.log" Nov 24 18:32:00 crc kubenswrapper[4808]: I1124 18:32:00.347374 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:32:00 crc kubenswrapper[4808]: E1124 18:32:00.347778 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:32:00 crc kubenswrapper[4808]: I1124 18:32:00.407318 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6mvpd_6f3367d5-54fd-4556-b1ba-5311e15433f3/speaker/0.log" Nov 24 18:32:00 crc kubenswrapper[4808]: I1124 18:32:00.459263 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b46nx_0e562199-e728-4698-8072-27b477593a5f/frr/0.log" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.347080 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:32:13 crc kubenswrapper[4808]: E1124 18:32:13.347704 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.556922 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp_432d4d4f-23e9-43f7-bf64-af0caa59dbea/util/0.log" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.670035 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7lr9s"] Nov 24 18:32:13 crc kubenswrapper[4808]: E1124 18:32:13.670408 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0244192-db95-42b9-9dd0-dd4d31f5d5e3" containerName="collect-profiles" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.670425 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0244192-db95-42b9-9dd0-dd4d31f5d5e3" containerName="collect-profiles" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.670657 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0244192-db95-42b9-9dd0-dd4d31f5d5e3" containerName="collect-profiles" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.672080 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.683628 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7lr9s"] Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.785960 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp_432d4d4f-23e9-43f7-bf64-af0caa59dbea/util/0.log" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.790717 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp_432d4d4f-23e9-43f7-bf64-af0caa59dbea/pull/0.log" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.813162 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g8n8\" (UniqueName: \"kubernetes.io/projected/d2f8138d-73a1-460c-b4b9-ac0422624a87-kube-api-access-5g8n8\") pod \"redhat-operators-7lr9s\" (UID: \"d2f8138d-73a1-460c-b4b9-ac0422624a87\") " pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.813204 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2f8138d-73a1-460c-b4b9-ac0422624a87-catalog-content\") pod \"redhat-operators-7lr9s\" (UID: \"d2f8138d-73a1-460c-b4b9-ac0422624a87\") " pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.813303 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2f8138d-73a1-460c-b4b9-ac0422624a87-utilities\") pod \"redhat-operators-7lr9s\" (UID: \"d2f8138d-73a1-460c-b4b9-ac0422624a87\") " pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.838731 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp_432d4d4f-23e9-43f7-bf64-af0caa59dbea/pull/0.log" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.915454 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g8n8\" (UniqueName: \"kubernetes.io/projected/d2f8138d-73a1-460c-b4b9-ac0422624a87-kube-api-access-5g8n8\") pod \"redhat-operators-7lr9s\" (UID: \"d2f8138d-73a1-460c-b4b9-ac0422624a87\") " pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.915501 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2f8138d-73a1-460c-b4b9-ac0422624a87-catalog-content\") pod \"redhat-operators-7lr9s\" (UID: \"d2f8138d-73a1-460c-b4b9-ac0422624a87\") " pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.915565 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2f8138d-73a1-460c-b4b9-ac0422624a87-utilities\") pod \"redhat-operators-7lr9s\" (UID: \"d2f8138d-73a1-460c-b4b9-ac0422624a87\") " pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.916119 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2f8138d-73a1-460c-b4b9-ac0422624a87-utilities\") pod \"redhat-operators-7lr9s\" (UID: \"d2f8138d-73a1-460c-b4b9-ac0422624a87\") " pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.916348 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2f8138d-73a1-460c-b4b9-ac0422624a87-catalog-content\") pod \"redhat-operators-7lr9s\" (UID: \"d2f8138d-73a1-460c-b4b9-ac0422624a87\") " pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.934635 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g8n8\" (UniqueName: \"kubernetes.io/projected/d2f8138d-73a1-460c-b4b9-ac0422624a87-kube-api-access-5g8n8\") pod \"redhat-operators-7lr9s\" (UID: \"d2f8138d-73a1-460c-b4b9-ac0422624a87\") " pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.992354 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp_432d4d4f-23e9-43f7-bf64-af0caa59dbea/pull/0.log" Nov 24 18:32:13 crc kubenswrapper[4808]: I1124 18:32:13.996631 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:14 crc kubenswrapper[4808]: I1124 18:32:14.071293 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp_432d4d4f-23e9-43f7-bf64-af0caa59dbea/extract/0.log" Nov 24 18:32:14 crc kubenswrapper[4808]: I1124 18:32:14.125646 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebzrpp_432d4d4f-23e9-43f7-bf64-af0caa59dbea/util/0.log" Nov 24 18:32:14 crc kubenswrapper[4808]: I1124 18:32:14.243199 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5j6bd_ba197cbd-9042-410f-adb4-d05cfec94ead/extract-utilities/0.log" Nov 24 18:32:14 crc kubenswrapper[4808]: I1124 18:32:14.424094 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5j6bd_ba197cbd-9042-410f-adb4-d05cfec94ead/extract-content/0.log" Nov 24 18:32:14 crc kubenswrapper[4808]: I1124 18:32:14.431065 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5j6bd_ba197cbd-9042-410f-adb4-d05cfec94ead/extract-content/0.log" Nov 24 18:32:14 crc kubenswrapper[4808]: I1124 18:32:14.495930 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5j6bd_ba197cbd-9042-410f-adb4-d05cfec94ead/extract-utilities/0.log" Nov 24 18:32:14 crc kubenswrapper[4808]: I1124 18:32:14.514187 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7lr9s"] Nov 24 18:32:14 crc kubenswrapper[4808]: I1124 18:32:14.694179 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5j6bd_ba197cbd-9042-410f-adb4-d05cfec94ead/extract-content/0.log" Nov 24 18:32:14 crc kubenswrapper[4808]: I1124 18:32:14.719968 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5j6bd_ba197cbd-9042-410f-adb4-d05cfec94ead/extract-utilities/0.log" Nov 24 18:32:14 crc kubenswrapper[4808]: I1124 18:32:14.728314 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7lr9s" event={"ID":"d2f8138d-73a1-460c-b4b9-ac0422624a87","Type":"ContainerStarted","Data":"0bb0f1123444065326f4ed678170e57402c7238173e6a5332cb02320d4fedd09"} Nov 24 18:32:15 crc kubenswrapper[4808]: I1124 18:32:15.022814 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rbql2_7929b2f8-b7af-4a85-9150-861fe438f602/extract-utilities/0.log" Nov 24 18:32:15 crc kubenswrapper[4808]: I1124 18:32:15.160098 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5j6bd_ba197cbd-9042-410f-adb4-d05cfec94ead/registry-server/0.log" Nov 24 18:32:15 crc kubenswrapper[4808]: I1124 18:32:15.284704 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rbql2_7929b2f8-b7af-4a85-9150-861fe438f602/extract-utilities/0.log" Nov 24 18:32:15 crc kubenswrapper[4808]: I1124 18:32:15.318205 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rbql2_7929b2f8-b7af-4a85-9150-861fe438f602/extract-content/0.log" Nov 24 18:32:15 crc kubenswrapper[4808]: I1124 18:32:15.320658 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rbql2_7929b2f8-b7af-4a85-9150-861fe438f602/extract-content/0.log" Nov 24 18:32:15 crc kubenswrapper[4808]: I1124 18:32:15.461905 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rbql2_7929b2f8-b7af-4a85-9150-861fe438f602/extract-utilities/0.log" Nov 24 18:32:15 crc kubenswrapper[4808]: I1124 18:32:15.491098 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rbql2_7929b2f8-b7af-4a85-9150-861fe438f602/extract-content/0.log" Nov 24 18:32:15 crc kubenswrapper[4808]: I1124 18:32:15.656946 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx_9d683a98-fa75-44d9-8d3d-11552b957864/util/0.log" Nov 24 18:32:15 crc kubenswrapper[4808]: I1124 18:32:15.739932 4808 generic.go:334] "Generic (PLEG): container finished" podID="d2f8138d-73a1-460c-b4b9-ac0422624a87" containerID="f45d5c350d7db5ee40a8ef822214e6aaa6e2b11d3f1412594d78bf63c308c47a" exitCode=0 Nov 24 18:32:15 crc kubenswrapper[4808]: I1124 18:32:15.739976 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7lr9s" event={"ID":"d2f8138d-73a1-460c-b4b9-ac0422624a87","Type":"ContainerDied","Data":"f45d5c350d7db5ee40a8ef822214e6aaa6e2b11d3f1412594d78bf63c308c47a"} Nov 24 18:32:15 crc kubenswrapper[4808]: I1124 18:32:15.742260 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 18:32:15 crc kubenswrapper[4808]: I1124 18:32:15.880125 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx_9d683a98-fa75-44d9-8d3d-11552b957864/util/0.log" Nov 24 18:32:15 crc kubenswrapper[4808]: I1124 18:32:15.913395 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx_9d683a98-fa75-44d9-8d3d-11552b957864/pull/0.log" Nov 24 18:32:15 crc kubenswrapper[4808]: I1124 18:32:15.939237 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx_9d683a98-fa75-44d9-8d3d-11552b957864/pull/0.log" Nov 24 18:32:16 crc kubenswrapper[4808]: I1124 18:32:16.051687 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rbql2_7929b2f8-b7af-4a85-9150-861fe438f602/registry-server/0.log" Nov 24 18:32:16 crc kubenswrapper[4808]: I1124 18:32:16.107363 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx_9d683a98-fa75-44d9-8d3d-11552b957864/pull/0.log" Nov 24 18:32:16 crc kubenswrapper[4808]: I1124 18:32:16.113443 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx_9d683a98-fa75-44d9-8d3d-11552b957864/util/0.log" Nov 24 18:32:16 crc kubenswrapper[4808]: I1124 18:32:16.143006 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c668wsx_9d683a98-fa75-44d9-8d3d-11552b957864/extract/0.log" Nov 24 18:32:16 crc kubenswrapper[4808]: I1124 18:32:16.264639 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-555rd_6ae01b29-6123-4219-b106-b8e55e83cb7f/marketplace-operator/0.log" Nov 24 18:32:16 crc kubenswrapper[4808]: I1124 18:32:16.332461 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pxdgf_14cc0e85-6bcb-443d-80f9-5f142532656a/extract-utilities/0.log" Nov 24 18:32:16 crc kubenswrapper[4808]: I1124 18:32:16.453859 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pxdgf_14cc0e85-6bcb-443d-80f9-5f142532656a/extract-content/0.log" Nov 24 18:32:16 crc kubenswrapper[4808]: I1124 18:32:16.483277 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pxdgf_14cc0e85-6bcb-443d-80f9-5f142532656a/extract-utilities/0.log" Nov 24 18:32:16 crc kubenswrapper[4808]: I1124 18:32:16.525399 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pxdgf_14cc0e85-6bcb-443d-80f9-5f142532656a/extract-content/0.log" Nov 24 18:32:16 crc kubenswrapper[4808]: I1124 18:32:16.665043 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pxdgf_14cc0e85-6bcb-443d-80f9-5f142532656a/extract-content/0.log" Nov 24 18:32:16 crc kubenswrapper[4808]: I1124 18:32:16.707269 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pxdgf_14cc0e85-6bcb-443d-80f9-5f142532656a/extract-utilities/0.log" Nov 24 18:32:16 crc kubenswrapper[4808]: I1124 18:32:16.748242 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7lr9s" event={"ID":"d2f8138d-73a1-460c-b4b9-ac0422624a87","Type":"ContainerStarted","Data":"0c9e61e8827a9e68649577bb4fdf5b2537a59f0a7c54801fc8798f909d0d6c2b"} Nov 24 18:32:16 crc kubenswrapper[4808]: I1124 18:32:16.813768 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pxdgf_14cc0e85-6bcb-443d-80f9-5f142532656a/registry-server/0.log" Nov 24 18:32:16 crc kubenswrapper[4808]: I1124 18:32:16.902860 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8kmz9_c13f73b0-c183-4f29-b75d-2fef31356a61/extract-utilities/0.log" Nov 24 18:32:17 crc kubenswrapper[4808]: I1124 18:32:17.039099 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8kmz9_c13f73b0-c183-4f29-b75d-2fef31356a61/extract-utilities/0.log" Nov 24 18:32:17 crc kubenswrapper[4808]: I1124 18:32:17.077789 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8kmz9_c13f73b0-c183-4f29-b75d-2fef31356a61/extract-content/0.log" Nov 24 18:32:17 crc kubenswrapper[4808]: I1124 18:32:17.092372 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8kmz9_c13f73b0-c183-4f29-b75d-2fef31356a61/extract-content/0.log" Nov 24 18:32:17 crc kubenswrapper[4808]: I1124 18:32:17.230130 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8kmz9_c13f73b0-c183-4f29-b75d-2fef31356a61/extract-utilities/0.log" Nov 24 18:32:17 crc kubenswrapper[4808]: I1124 18:32:17.368385 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8kmz9_c13f73b0-c183-4f29-b75d-2fef31356a61/extract-content/0.log" Nov 24 18:32:17 crc kubenswrapper[4808]: I1124 18:32:17.698883 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8kmz9_c13f73b0-c183-4f29-b75d-2fef31356a61/registry-server/0.log" Nov 24 18:32:17 crc kubenswrapper[4808]: I1124 18:32:17.760353 4808 generic.go:334] "Generic (PLEG): container finished" podID="d2f8138d-73a1-460c-b4b9-ac0422624a87" containerID="0c9e61e8827a9e68649577bb4fdf5b2537a59f0a7c54801fc8798f909d0d6c2b" exitCode=0 Nov 24 18:32:17 crc kubenswrapper[4808]: I1124 18:32:17.760566 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7lr9s" event={"ID":"d2f8138d-73a1-460c-b4b9-ac0422624a87","Type":"ContainerDied","Data":"0c9e61e8827a9e68649577bb4fdf5b2537a59f0a7c54801fc8798f909d0d6c2b"} Nov 24 18:32:18 crc kubenswrapper[4808]: I1124 18:32:18.774516 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7lr9s" event={"ID":"d2f8138d-73a1-460c-b4b9-ac0422624a87","Type":"ContainerStarted","Data":"0beb68b3bd8cbf57d318a39396581d77c9617bfbbf9b0d483ba5a6913178c363"} Nov 24 18:32:23 crc kubenswrapper[4808]: I1124 18:32:23.997396 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:23 crc kubenswrapper[4808]: I1124 18:32:23.998108 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:25 crc kubenswrapper[4808]: I1124 18:32:25.057398 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7lr9s" podUID="d2f8138d-73a1-460c-b4b9-ac0422624a87" containerName="registry-server" probeResult="failure" output=< Nov 24 18:32:25 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Nov 24 18:32:25 crc kubenswrapper[4808]: > Nov 24 18:32:26 crc kubenswrapper[4808]: I1124 18:32:26.348522 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:32:26 crc kubenswrapper[4808]: E1124 18:32:26.348840 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:32:34 crc kubenswrapper[4808]: I1124 18:32:34.049661 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:34 crc kubenswrapper[4808]: I1124 18:32:34.079086 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7lr9s" podStartSLOduration=18.607489961 podStartE2EDuration="21.079065487s" podCreationTimestamp="2025-11-24 18:32:13 +0000 UTC" firstStartedPulling="2025-11-24 18:32:15.742040256 +0000 UTC m=+3928.339708058" lastFinishedPulling="2025-11-24 18:32:18.213615782 +0000 UTC m=+3930.811283584" observedRunningTime="2025-11-24 18:32:18.797975533 +0000 UTC m=+3931.395643345" watchObservedRunningTime="2025-11-24 18:32:34.079065487 +0000 UTC m=+3946.676733289" Nov 24 18:32:34 crc kubenswrapper[4808]: I1124 18:32:34.124919 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:34 crc kubenswrapper[4808]: I1124 18:32:34.298498 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7lr9s"] Nov 24 18:32:35 crc kubenswrapper[4808]: I1124 18:32:35.938578 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7lr9s" podUID="d2f8138d-73a1-460c-b4b9-ac0422624a87" containerName="registry-server" containerID="cri-o://0beb68b3bd8cbf57d318a39396581d77c9617bfbbf9b0d483ba5a6913178c363" gracePeriod=2 Nov 24 18:32:36 crc kubenswrapper[4808]: I1124 18:32:36.424451 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:36 crc kubenswrapper[4808]: I1124 18:32:36.542435 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g8n8\" (UniqueName: \"kubernetes.io/projected/d2f8138d-73a1-460c-b4b9-ac0422624a87-kube-api-access-5g8n8\") pod \"d2f8138d-73a1-460c-b4b9-ac0422624a87\" (UID: \"d2f8138d-73a1-460c-b4b9-ac0422624a87\") " Nov 24 18:32:36 crc kubenswrapper[4808]: I1124 18:32:36.542725 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2f8138d-73a1-460c-b4b9-ac0422624a87-catalog-content\") pod \"d2f8138d-73a1-460c-b4b9-ac0422624a87\" (UID: \"d2f8138d-73a1-460c-b4b9-ac0422624a87\") " Nov 24 18:32:36 crc kubenswrapper[4808]: I1124 18:32:36.542786 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2f8138d-73a1-460c-b4b9-ac0422624a87-utilities\") pod \"d2f8138d-73a1-460c-b4b9-ac0422624a87\" (UID: \"d2f8138d-73a1-460c-b4b9-ac0422624a87\") " Nov 24 18:32:36 crc kubenswrapper[4808]: I1124 18:32:36.543721 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2f8138d-73a1-460c-b4b9-ac0422624a87-utilities" (OuterVolumeSpecName: "utilities") pod "d2f8138d-73a1-460c-b4b9-ac0422624a87" (UID: "d2f8138d-73a1-460c-b4b9-ac0422624a87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:32:36 crc kubenswrapper[4808]: I1124 18:32:36.548671 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2f8138d-73a1-460c-b4b9-ac0422624a87-kube-api-access-5g8n8" (OuterVolumeSpecName: "kube-api-access-5g8n8") pod "d2f8138d-73a1-460c-b4b9-ac0422624a87" (UID: "d2f8138d-73a1-460c-b4b9-ac0422624a87"). InnerVolumeSpecName "kube-api-access-5g8n8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:32:36 crc kubenswrapper[4808]: I1124 18:32:36.645457 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g8n8\" (UniqueName: \"kubernetes.io/projected/d2f8138d-73a1-460c-b4b9-ac0422624a87-kube-api-access-5g8n8\") on node \"crc\" DevicePath \"\"" Nov 24 18:32:36 crc kubenswrapper[4808]: I1124 18:32:36.645489 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2f8138d-73a1-460c-b4b9-ac0422624a87-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:32:36 crc kubenswrapper[4808]: I1124 18:32:36.651920 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2f8138d-73a1-460c-b4b9-ac0422624a87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2f8138d-73a1-460c-b4b9-ac0422624a87" (UID: "d2f8138d-73a1-460c-b4b9-ac0422624a87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:32:36 crc kubenswrapper[4808]: I1124 18:32:36.746408 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2f8138d-73a1-460c-b4b9-ac0422624a87-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:32:36 crc kubenswrapper[4808]: I1124 18:32:36.947841 4808 generic.go:334] "Generic (PLEG): container finished" podID="d2f8138d-73a1-460c-b4b9-ac0422624a87" containerID="0beb68b3bd8cbf57d318a39396581d77c9617bfbbf9b0d483ba5a6913178c363" exitCode=0 Nov 24 18:32:36 crc kubenswrapper[4808]: I1124 18:32:36.947884 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7lr9s" event={"ID":"d2f8138d-73a1-460c-b4b9-ac0422624a87","Type":"ContainerDied","Data":"0beb68b3bd8cbf57d318a39396581d77c9617bfbbf9b0d483ba5a6913178c363"} Nov 24 18:32:36 crc kubenswrapper[4808]: I1124 18:32:36.947914 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7lr9s" event={"ID":"d2f8138d-73a1-460c-b4b9-ac0422624a87","Type":"ContainerDied","Data":"0bb0f1123444065326f4ed678170e57402c7238173e6a5332cb02320d4fedd09"} Nov 24 18:32:36 crc kubenswrapper[4808]: I1124 18:32:36.947933 4808 scope.go:117] "RemoveContainer" containerID="0beb68b3bd8cbf57d318a39396581d77c9617bfbbf9b0d483ba5a6913178c363" Nov 24 18:32:36 crc kubenswrapper[4808]: I1124 18:32:36.948080 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7lr9s" Nov 24 18:32:37 crc kubenswrapper[4808]: I1124 18:32:37.001236 4808 scope.go:117] "RemoveContainer" containerID="0c9e61e8827a9e68649577bb4fdf5b2537a59f0a7c54801fc8798f909d0d6c2b" Nov 24 18:32:37 crc kubenswrapper[4808]: I1124 18:32:37.011064 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7lr9s"] Nov 24 18:32:37 crc kubenswrapper[4808]: I1124 18:32:37.017722 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7lr9s"] Nov 24 18:32:37 crc kubenswrapper[4808]: I1124 18:32:37.019938 4808 scope.go:117] "RemoveContainer" containerID="f45d5c350d7db5ee40a8ef822214e6aaa6e2b11d3f1412594d78bf63c308c47a" Nov 24 18:32:37 crc kubenswrapper[4808]: I1124 18:32:37.094052 4808 scope.go:117] "RemoveContainer" containerID="0beb68b3bd8cbf57d318a39396581d77c9617bfbbf9b0d483ba5a6913178c363" Nov 24 18:32:37 crc kubenswrapper[4808]: E1124 18:32:37.097587 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0beb68b3bd8cbf57d318a39396581d77c9617bfbbf9b0d483ba5a6913178c363\": container with ID starting with 0beb68b3bd8cbf57d318a39396581d77c9617bfbbf9b0d483ba5a6913178c363 not found: ID does not exist" containerID="0beb68b3bd8cbf57d318a39396581d77c9617bfbbf9b0d483ba5a6913178c363" Nov 24 18:32:37 crc kubenswrapper[4808]: I1124 18:32:37.097623 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0beb68b3bd8cbf57d318a39396581d77c9617bfbbf9b0d483ba5a6913178c363"} err="failed to get container status \"0beb68b3bd8cbf57d318a39396581d77c9617bfbbf9b0d483ba5a6913178c363\": rpc error: code = NotFound desc = could not find container \"0beb68b3bd8cbf57d318a39396581d77c9617bfbbf9b0d483ba5a6913178c363\": container with ID starting with 0beb68b3bd8cbf57d318a39396581d77c9617bfbbf9b0d483ba5a6913178c363 not found: ID does not exist" Nov 24 18:32:37 crc kubenswrapper[4808]: I1124 18:32:37.097643 4808 scope.go:117] "RemoveContainer" containerID="0c9e61e8827a9e68649577bb4fdf5b2537a59f0a7c54801fc8798f909d0d6c2b" Nov 24 18:32:37 crc kubenswrapper[4808]: E1124 18:32:37.101095 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c9e61e8827a9e68649577bb4fdf5b2537a59f0a7c54801fc8798f909d0d6c2b\": container with ID starting with 0c9e61e8827a9e68649577bb4fdf5b2537a59f0a7c54801fc8798f909d0d6c2b not found: ID does not exist" containerID="0c9e61e8827a9e68649577bb4fdf5b2537a59f0a7c54801fc8798f909d0d6c2b" Nov 24 18:32:37 crc kubenswrapper[4808]: I1124 18:32:37.101124 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c9e61e8827a9e68649577bb4fdf5b2537a59f0a7c54801fc8798f909d0d6c2b"} err="failed to get container status \"0c9e61e8827a9e68649577bb4fdf5b2537a59f0a7c54801fc8798f909d0d6c2b\": rpc error: code = NotFound desc = could not find container \"0c9e61e8827a9e68649577bb4fdf5b2537a59f0a7c54801fc8798f909d0d6c2b\": container with ID starting with 0c9e61e8827a9e68649577bb4fdf5b2537a59f0a7c54801fc8798f909d0d6c2b not found: ID does not exist" Nov 24 18:32:37 crc kubenswrapper[4808]: I1124 18:32:37.101139 4808 scope.go:117] "RemoveContainer" containerID="f45d5c350d7db5ee40a8ef822214e6aaa6e2b11d3f1412594d78bf63c308c47a" Nov 24 18:32:37 crc kubenswrapper[4808]: E1124 18:32:37.105114 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f45d5c350d7db5ee40a8ef822214e6aaa6e2b11d3f1412594d78bf63c308c47a\": container with ID starting with f45d5c350d7db5ee40a8ef822214e6aaa6e2b11d3f1412594d78bf63c308c47a not found: ID does not exist" containerID="f45d5c350d7db5ee40a8ef822214e6aaa6e2b11d3f1412594d78bf63c308c47a" Nov 24 18:32:37 crc kubenswrapper[4808]: I1124 18:32:37.105143 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f45d5c350d7db5ee40a8ef822214e6aaa6e2b11d3f1412594d78bf63c308c47a"} err="failed to get container status \"f45d5c350d7db5ee40a8ef822214e6aaa6e2b11d3f1412594d78bf63c308c47a\": rpc error: code = NotFound desc = could not find container \"f45d5c350d7db5ee40a8ef822214e6aaa6e2b11d3f1412594d78bf63c308c47a\": container with ID starting with f45d5c350d7db5ee40a8ef822214e6aaa6e2b11d3f1412594d78bf63c308c47a not found: ID does not exist" Nov 24 18:32:38 crc kubenswrapper[4808]: I1124 18:32:38.359958 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2f8138d-73a1-460c-b4b9-ac0422624a87" path="/var/lib/kubelet/pods/d2f8138d-73a1-460c-b4b9-ac0422624a87/volumes" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.102865 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zx9hk"] Nov 24 18:32:41 crc kubenswrapper[4808]: E1124 18:32:41.103687 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2f8138d-73a1-460c-b4b9-ac0422624a87" containerName="extract-content" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.103701 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2f8138d-73a1-460c-b4b9-ac0422624a87" containerName="extract-content" Nov 24 18:32:41 crc kubenswrapper[4808]: E1124 18:32:41.103721 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2f8138d-73a1-460c-b4b9-ac0422624a87" containerName="extract-utilities" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.103727 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2f8138d-73a1-460c-b4b9-ac0422624a87" containerName="extract-utilities" Nov 24 18:32:41 crc kubenswrapper[4808]: E1124 18:32:41.103742 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2f8138d-73a1-460c-b4b9-ac0422624a87" containerName="registry-server" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.103749 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2f8138d-73a1-460c-b4b9-ac0422624a87" containerName="registry-server" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.103952 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2f8138d-73a1-460c-b4b9-ac0422624a87" containerName="registry-server" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.105302 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.129183 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zx9hk"] Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.237278 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkhkk\" (UniqueName: \"kubernetes.io/projected/a4503f95-37ef-4564-9038-895faaa8d07d-kube-api-access-nkhkk\") pod \"community-operators-zx9hk\" (UID: \"a4503f95-37ef-4564-9038-895faaa8d07d\") " pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.237555 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4503f95-37ef-4564-9038-895faaa8d07d-catalog-content\") pod \"community-operators-zx9hk\" (UID: \"a4503f95-37ef-4564-9038-895faaa8d07d\") " pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.237750 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4503f95-37ef-4564-9038-895faaa8d07d-utilities\") pod \"community-operators-zx9hk\" (UID: \"a4503f95-37ef-4564-9038-895faaa8d07d\") " pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.339028 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4503f95-37ef-4564-9038-895faaa8d07d-utilities\") pod \"community-operators-zx9hk\" (UID: \"a4503f95-37ef-4564-9038-895faaa8d07d\") " pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.339121 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkhkk\" (UniqueName: \"kubernetes.io/projected/a4503f95-37ef-4564-9038-895faaa8d07d-kube-api-access-nkhkk\") pod \"community-operators-zx9hk\" (UID: \"a4503f95-37ef-4564-9038-895faaa8d07d\") " pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.339203 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4503f95-37ef-4564-9038-895faaa8d07d-catalog-content\") pod \"community-operators-zx9hk\" (UID: \"a4503f95-37ef-4564-9038-895faaa8d07d\") " pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.339525 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4503f95-37ef-4564-9038-895faaa8d07d-utilities\") pod \"community-operators-zx9hk\" (UID: \"a4503f95-37ef-4564-9038-895faaa8d07d\") " pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.339598 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4503f95-37ef-4564-9038-895faaa8d07d-catalog-content\") pod \"community-operators-zx9hk\" (UID: \"a4503f95-37ef-4564-9038-895faaa8d07d\") " pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.347542 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:32:41 crc kubenswrapper[4808]: E1124 18:32:41.347756 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.363521 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkhkk\" (UniqueName: \"kubernetes.io/projected/a4503f95-37ef-4564-9038-895faaa8d07d-kube-api-access-nkhkk\") pod \"community-operators-zx9hk\" (UID: \"a4503f95-37ef-4564-9038-895faaa8d07d\") " pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:41 crc kubenswrapper[4808]: I1124 18:32:41.421401 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:42 crc kubenswrapper[4808]: I1124 18:32:42.040562 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zx9hk"] Nov 24 18:32:43 crc kubenswrapper[4808]: I1124 18:32:43.011629 4808 generic.go:334] "Generic (PLEG): container finished" podID="a4503f95-37ef-4564-9038-895faaa8d07d" containerID="9f1875e62b4e98aab2819c95ad7cd1c177434e2a86dc357e8fec6143c288cc91" exitCode=0 Nov 24 18:32:43 crc kubenswrapper[4808]: I1124 18:32:43.011814 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx9hk" event={"ID":"a4503f95-37ef-4564-9038-895faaa8d07d","Type":"ContainerDied","Data":"9f1875e62b4e98aab2819c95ad7cd1c177434e2a86dc357e8fec6143c288cc91"} Nov 24 18:32:43 crc kubenswrapper[4808]: I1124 18:32:43.012070 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx9hk" event={"ID":"a4503f95-37ef-4564-9038-895faaa8d07d","Type":"ContainerStarted","Data":"23c341c2e02153e6033c393e2b30f1b3aeb8aee7cd907f9669a7b05ade2670a1"} Nov 24 18:32:44 crc kubenswrapper[4808]: I1124 18:32:44.021340 4808 generic.go:334] "Generic (PLEG): container finished" podID="a4503f95-37ef-4564-9038-895faaa8d07d" containerID="3e464173b53738df6610fee3f6a4af0ba24b5f2dddd49b07525d5edff28afadb" exitCode=0 Nov 24 18:32:44 crc kubenswrapper[4808]: I1124 18:32:44.021447 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx9hk" event={"ID":"a4503f95-37ef-4564-9038-895faaa8d07d","Type":"ContainerDied","Data":"3e464173b53738df6610fee3f6a4af0ba24b5f2dddd49b07525d5edff28afadb"} Nov 24 18:32:45 crc kubenswrapper[4808]: I1124 18:32:45.034553 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx9hk" event={"ID":"a4503f95-37ef-4564-9038-895faaa8d07d","Type":"ContainerStarted","Data":"124188be9731ef502ea6aab2569bd6f1dc636f38b8c9f15496f785786105a87b"} Nov 24 18:32:45 crc kubenswrapper[4808]: I1124 18:32:45.059982 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zx9hk" podStartSLOduration=2.630400311 podStartE2EDuration="4.059961576s" podCreationTimestamp="2025-11-24 18:32:41 +0000 UTC" firstStartedPulling="2025-11-24 18:32:43.013703352 +0000 UTC m=+3955.611371154" lastFinishedPulling="2025-11-24 18:32:44.443264617 +0000 UTC m=+3957.040932419" observedRunningTime="2025-11-24 18:32:45.052205272 +0000 UTC m=+3957.649873074" watchObservedRunningTime="2025-11-24 18:32:45.059961576 +0000 UTC m=+3957.657629378" Nov 24 18:32:48 crc kubenswrapper[4808]: E1124 18:32:48.964031 4808 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.171:58260->38.129.56.171:34569: write tcp 38.129.56.171:58260->38.129.56.171:34569: write: broken pipe Nov 24 18:32:51 crc kubenswrapper[4808]: I1124 18:32:51.422504 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:51 crc kubenswrapper[4808]: I1124 18:32:51.422986 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:52 crc kubenswrapper[4808]: I1124 18:32:52.409091 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:52 crc kubenswrapper[4808]: I1124 18:32:52.492944 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:52 crc kubenswrapper[4808]: I1124 18:32:52.667096 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zx9hk"] Nov 24 18:32:54 crc kubenswrapper[4808]: I1124 18:32:54.147896 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zx9hk" podUID="a4503f95-37ef-4564-9038-895faaa8d07d" containerName="registry-server" containerID="cri-o://124188be9731ef502ea6aab2569bd6f1dc636f38b8c9f15496f785786105a87b" gracePeriod=2 Nov 24 18:32:54 crc kubenswrapper[4808]: I1124 18:32:54.666779 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:54 crc kubenswrapper[4808]: I1124 18:32:54.775133 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkhkk\" (UniqueName: \"kubernetes.io/projected/a4503f95-37ef-4564-9038-895faaa8d07d-kube-api-access-nkhkk\") pod \"a4503f95-37ef-4564-9038-895faaa8d07d\" (UID: \"a4503f95-37ef-4564-9038-895faaa8d07d\") " Nov 24 18:32:54 crc kubenswrapper[4808]: I1124 18:32:54.775230 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4503f95-37ef-4564-9038-895faaa8d07d-catalog-content\") pod \"a4503f95-37ef-4564-9038-895faaa8d07d\" (UID: \"a4503f95-37ef-4564-9038-895faaa8d07d\") " Nov 24 18:32:54 crc kubenswrapper[4808]: I1124 18:32:54.775312 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4503f95-37ef-4564-9038-895faaa8d07d-utilities\") pod \"a4503f95-37ef-4564-9038-895faaa8d07d\" (UID: \"a4503f95-37ef-4564-9038-895faaa8d07d\") " Nov 24 18:32:54 crc kubenswrapper[4808]: I1124 18:32:54.776299 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4503f95-37ef-4564-9038-895faaa8d07d-utilities" (OuterVolumeSpecName: "utilities") pod "a4503f95-37ef-4564-9038-895faaa8d07d" (UID: "a4503f95-37ef-4564-9038-895faaa8d07d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:32:54 crc kubenswrapper[4808]: I1124 18:32:54.784810 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4503f95-37ef-4564-9038-895faaa8d07d-kube-api-access-nkhkk" (OuterVolumeSpecName: "kube-api-access-nkhkk") pod "a4503f95-37ef-4564-9038-895faaa8d07d" (UID: "a4503f95-37ef-4564-9038-895faaa8d07d"). InnerVolumeSpecName "kube-api-access-nkhkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:32:54 crc kubenswrapper[4808]: I1124 18:32:54.855261 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4503f95-37ef-4564-9038-895faaa8d07d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4503f95-37ef-4564-9038-895faaa8d07d" (UID: "a4503f95-37ef-4564-9038-895faaa8d07d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:32:54 crc kubenswrapper[4808]: I1124 18:32:54.877845 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4503f95-37ef-4564-9038-895faaa8d07d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:32:54 crc kubenswrapper[4808]: I1124 18:32:54.877893 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4503f95-37ef-4564-9038-895faaa8d07d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:32:54 crc kubenswrapper[4808]: I1124 18:32:54.877910 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkhkk\" (UniqueName: \"kubernetes.io/projected/a4503f95-37ef-4564-9038-895faaa8d07d-kube-api-access-nkhkk\") on node \"crc\" DevicePath \"\"" Nov 24 18:32:55 crc kubenswrapper[4808]: I1124 18:32:55.164417 4808 generic.go:334] "Generic (PLEG): container finished" podID="a4503f95-37ef-4564-9038-895faaa8d07d" containerID="124188be9731ef502ea6aab2569bd6f1dc636f38b8c9f15496f785786105a87b" exitCode=0 Nov 24 18:32:55 crc kubenswrapper[4808]: I1124 18:32:55.164475 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx9hk" event={"ID":"a4503f95-37ef-4564-9038-895faaa8d07d","Type":"ContainerDied","Data":"124188be9731ef502ea6aab2569bd6f1dc636f38b8c9f15496f785786105a87b"} Nov 24 18:32:55 crc kubenswrapper[4808]: I1124 18:32:55.164859 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx9hk" event={"ID":"a4503f95-37ef-4564-9038-895faaa8d07d","Type":"ContainerDied","Data":"23c341c2e02153e6033c393e2b30f1b3aeb8aee7cd907f9669a7b05ade2670a1"} Nov 24 18:32:55 crc kubenswrapper[4808]: I1124 18:32:55.164886 4808 scope.go:117] "RemoveContainer" containerID="124188be9731ef502ea6aab2569bd6f1dc636f38b8c9f15496f785786105a87b" Nov 24 18:32:55 crc kubenswrapper[4808]: I1124 18:32:55.164516 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx9hk" Nov 24 18:32:55 crc kubenswrapper[4808]: I1124 18:32:55.205300 4808 scope.go:117] "RemoveContainer" containerID="3e464173b53738df6610fee3f6a4af0ba24b5f2dddd49b07525d5edff28afadb" Nov 24 18:32:55 crc kubenswrapper[4808]: I1124 18:32:55.213927 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zx9hk"] Nov 24 18:32:55 crc kubenswrapper[4808]: I1124 18:32:55.223409 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zx9hk"] Nov 24 18:32:55 crc kubenswrapper[4808]: I1124 18:32:55.245598 4808 scope.go:117] "RemoveContainer" containerID="9f1875e62b4e98aab2819c95ad7cd1c177434e2a86dc357e8fec6143c288cc91" Nov 24 18:32:55 crc kubenswrapper[4808]: I1124 18:32:55.302961 4808 scope.go:117] "RemoveContainer" containerID="124188be9731ef502ea6aab2569bd6f1dc636f38b8c9f15496f785786105a87b" Nov 24 18:32:55 crc kubenswrapper[4808]: E1124 18:32:55.303993 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"124188be9731ef502ea6aab2569bd6f1dc636f38b8c9f15496f785786105a87b\": container with ID starting with 124188be9731ef502ea6aab2569bd6f1dc636f38b8c9f15496f785786105a87b not found: ID does not exist" containerID="124188be9731ef502ea6aab2569bd6f1dc636f38b8c9f15496f785786105a87b" Nov 24 18:32:55 crc kubenswrapper[4808]: I1124 18:32:55.304109 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"124188be9731ef502ea6aab2569bd6f1dc636f38b8c9f15496f785786105a87b"} err="failed to get container status \"124188be9731ef502ea6aab2569bd6f1dc636f38b8c9f15496f785786105a87b\": rpc error: code = NotFound desc = could not find container \"124188be9731ef502ea6aab2569bd6f1dc636f38b8c9f15496f785786105a87b\": container with ID starting with 124188be9731ef502ea6aab2569bd6f1dc636f38b8c9f15496f785786105a87b not found: ID does not exist" Nov 24 18:32:55 crc kubenswrapper[4808]: I1124 18:32:55.304192 4808 scope.go:117] "RemoveContainer" containerID="3e464173b53738df6610fee3f6a4af0ba24b5f2dddd49b07525d5edff28afadb" Nov 24 18:32:55 crc kubenswrapper[4808]: E1124 18:32:55.304833 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e464173b53738df6610fee3f6a4af0ba24b5f2dddd49b07525d5edff28afadb\": container with ID starting with 3e464173b53738df6610fee3f6a4af0ba24b5f2dddd49b07525d5edff28afadb not found: ID does not exist" containerID="3e464173b53738df6610fee3f6a4af0ba24b5f2dddd49b07525d5edff28afadb" Nov 24 18:32:55 crc kubenswrapper[4808]: I1124 18:32:55.304885 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e464173b53738df6610fee3f6a4af0ba24b5f2dddd49b07525d5edff28afadb"} err="failed to get container status \"3e464173b53738df6610fee3f6a4af0ba24b5f2dddd49b07525d5edff28afadb\": rpc error: code = NotFound desc = could not find container \"3e464173b53738df6610fee3f6a4af0ba24b5f2dddd49b07525d5edff28afadb\": container with ID starting with 3e464173b53738df6610fee3f6a4af0ba24b5f2dddd49b07525d5edff28afadb not found: ID does not exist" Nov 24 18:32:55 crc kubenswrapper[4808]: I1124 18:32:55.304920 4808 scope.go:117] "RemoveContainer" containerID="9f1875e62b4e98aab2819c95ad7cd1c177434e2a86dc357e8fec6143c288cc91" Nov 24 18:32:55 crc kubenswrapper[4808]: E1124 18:32:55.305456 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f1875e62b4e98aab2819c95ad7cd1c177434e2a86dc357e8fec6143c288cc91\": container with ID starting with 9f1875e62b4e98aab2819c95ad7cd1c177434e2a86dc357e8fec6143c288cc91 not found: ID does not exist" containerID="9f1875e62b4e98aab2819c95ad7cd1c177434e2a86dc357e8fec6143c288cc91" Nov 24 18:32:55 crc kubenswrapper[4808]: I1124 18:32:55.305493 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f1875e62b4e98aab2819c95ad7cd1c177434e2a86dc357e8fec6143c288cc91"} err="failed to get container status \"9f1875e62b4e98aab2819c95ad7cd1c177434e2a86dc357e8fec6143c288cc91\": rpc error: code = NotFound desc = could not find container \"9f1875e62b4e98aab2819c95ad7cd1c177434e2a86dc357e8fec6143c288cc91\": container with ID starting with 9f1875e62b4e98aab2819c95ad7cd1c177434e2a86dc357e8fec6143c288cc91 not found: ID does not exist" Nov 24 18:32:56 crc kubenswrapper[4808]: I1124 18:32:56.347917 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:32:56 crc kubenswrapper[4808]: E1124 18:32:56.348384 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:32:56 crc kubenswrapper[4808]: I1124 18:32:56.359412 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4503f95-37ef-4564-9038-895faaa8d07d" path="/var/lib/kubelet/pods/a4503f95-37ef-4564-9038-895faaa8d07d/volumes" Nov 24 18:33:07 crc kubenswrapper[4808]: I1124 18:33:07.348676 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:33:07 crc kubenswrapper[4808]: E1124 18:33:07.351813 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:33:21 crc kubenswrapper[4808]: I1124 18:33:21.347911 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:33:21 crc kubenswrapper[4808]: E1124 18:33:21.348937 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:33:34 crc kubenswrapper[4808]: I1124 18:33:34.360235 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:33:34 crc kubenswrapper[4808]: E1124 18:33:34.362815 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ht87w_openshift-machine-config-operator(5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" Nov 24 18:33:47 crc kubenswrapper[4808]: I1124 18:33:47.347094 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:33:47 crc kubenswrapper[4808]: I1124 18:33:47.817586 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"edaf6fa37eb97fe11a3c54d186e8fc013d121dd3e78f8413570dc6ca074c8458"} Nov 24 18:33:54 crc kubenswrapper[4808]: I1124 18:33:54.920265 4808 generic.go:334] "Generic (PLEG): container finished" podID="626e8db0-9921-43ac-a1e1-55ff96937a43" containerID="71bdc92f32291232187da14655de079d7bce31d4f354b1c4ee44415bbcba1bf2" exitCode=0 Nov 24 18:33:54 crc kubenswrapper[4808]: I1124 18:33:54.920365 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2lp6/must-gather-hjrb8" event={"ID":"626e8db0-9921-43ac-a1e1-55ff96937a43","Type":"ContainerDied","Data":"71bdc92f32291232187da14655de079d7bce31d4f354b1c4ee44415bbcba1bf2"} Nov 24 18:33:54 crc kubenswrapper[4808]: I1124 18:33:54.921952 4808 scope.go:117] "RemoveContainer" containerID="71bdc92f32291232187da14655de079d7bce31d4f354b1c4ee44415bbcba1bf2" Nov 24 18:33:55 crc kubenswrapper[4808]: I1124 18:33:55.515891 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-d2lp6_must-gather-hjrb8_626e8db0-9921-43ac-a1e1-55ff96937a43/gather/0.log" Nov 24 18:34:05 crc kubenswrapper[4808]: I1124 18:34:05.382758 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-d2lp6/must-gather-hjrb8"] Nov 24 18:34:05 crc kubenswrapper[4808]: I1124 18:34:05.383713 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-d2lp6/must-gather-hjrb8" podUID="626e8db0-9921-43ac-a1e1-55ff96937a43" containerName="copy" containerID="cri-o://f90518d8aa5aad6b9761866d87910ff71bac9813edb2c684c3570b4efe803eb1" gracePeriod=2 Nov 24 18:34:05 crc kubenswrapper[4808]: I1124 18:34:05.395125 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-d2lp6/must-gather-hjrb8"] Nov 24 18:34:05 crc kubenswrapper[4808]: I1124 18:34:05.844674 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-d2lp6_must-gather-hjrb8_626e8db0-9921-43ac-a1e1-55ff96937a43/copy/0.log" Nov 24 18:34:05 crc kubenswrapper[4808]: I1124 18:34:05.845403 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/must-gather-hjrb8" Nov 24 18:34:05 crc kubenswrapper[4808]: I1124 18:34:05.965694 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88rdg\" (UniqueName: \"kubernetes.io/projected/626e8db0-9921-43ac-a1e1-55ff96937a43-kube-api-access-88rdg\") pod \"626e8db0-9921-43ac-a1e1-55ff96937a43\" (UID: \"626e8db0-9921-43ac-a1e1-55ff96937a43\") " Nov 24 18:34:05 crc kubenswrapper[4808]: I1124 18:34:05.965755 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/626e8db0-9921-43ac-a1e1-55ff96937a43-must-gather-output\") pod \"626e8db0-9921-43ac-a1e1-55ff96937a43\" (UID: \"626e8db0-9921-43ac-a1e1-55ff96937a43\") " Nov 24 18:34:05 crc kubenswrapper[4808]: I1124 18:34:05.981784 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/626e8db0-9921-43ac-a1e1-55ff96937a43-kube-api-access-88rdg" (OuterVolumeSpecName: "kube-api-access-88rdg") pod "626e8db0-9921-43ac-a1e1-55ff96937a43" (UID: "626e8db0-9921-43ac-a1e1-55ff96937a43"). InnerVolumeSpecName "kube-api-access-88rdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:34:06 crc kubenswrapper[4808]: I1124 18:34:06.040602 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-d2lp6_must-gather-hjrb8_626e8db0-9921-43ac-a1e1-55ff96937a43/copy/0.log" Nov 24 18:34:06 crc kubenswrapper[4808]: I1124 18:34:06.041152 4808 generic.go:334] "Generic (PLEG): container finished" podID="626e8db0-9921-43ac-a1e1-55ff96937a43" containerID="f90518d8aa5aad6b9761866d87910ff71bac9813edb2c684c3570b4efe803eb1" exitCode=143 Nov 24 18:34:06 crc kubenswrapper[4808]: I1124 18:34:06.041203 4808 scope.go:117] "RemoveContainer" containerID="f90518d8aa5aad6b9761866d87910ff71bac9813edb2c684c3570b4efe803eb1" Nov 24 18:34:06 crc kubenswrapper[4808]: I1124 18:34:06.041241 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2lp6/must-gather-hjrb8" Nov 24 18:34:06 crc kubenswrapper[4808]: I1124 18:34:06.057449 4808 scope.go:117] "RemoveContainer" containerID="71bdc92f32291232187da14655de079d7bce31d4f354b1c4ee44415bbcba1bf2" Nov 24 18:34:06 crc kubenswrapper[4808]: I1124 18:34:06.074172 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88rdg\" (UniqueName: \"kubernetes.io/projected/626e8db0-9921-43ac-a1e1-55ff96937a43-kube-api-access-88rdg\") on node \"crc\" DevicePath \"\"" Nov 24 18:34:06 crc kubenswrapper[4808]: I1124 18:34:06.100796 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/626e8db0-9921-43ac-a1e1-55ff96937a43-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "626e8db0-9921-43ac-a1e1-55ff96937a43" (UID: "626e8db0-9921-43ac-a1e1-55ff96937a43"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:34:06 crc kubenswrapper[4808]: I1124 18:34:06.131058 4808 scope.go:117] "RemoveContainer" containerID="f90518d8aa5aad6b9761866d87910ff71bac9813edb2c684c3570b4efe803eb1" Nov 24 18:34:06 crc kubenswrapper[4808]: E1124 18:34:06.132045 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f90518d8aa5aad6b9761866d87910ff71bac9813edb2c684c3570b4efe803eb1\": container with ID starting with f90518d8aa5aad6b9761866d87910ff71bac9813edb2c684c3570b4efe803eb1 not found: ID does not exist" containerID="f90518d8aa5aad6b9761866d87910ff71bac9813edb2c684c3570b4efe803eb1" Nov 24 18:34:06 crc kubenswrapper[4808]: I1124 18:34:06.133331 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f90518d8aa5aad6b9761866d87910ff71bac9813edb2c684c3570b4efe803eb1"} err="failed to get container status \"f90518d8aa5aad6b9761866d87910ff71bac9813edb2c684c3570b4efe803eb1\": rpc error: code = NotFound desc = could not find container \"f90518d8aa5aad6b9761866d87910ff71bac9813edb2c684c3570b4efe803eb1\": container with ID starting with f90518d8aa5aad6b9761866d87910ff71bac9813edb2c684c3570b4efe803eb1 not found: ID does not exist" Nov 24 18:34:06 crc kubenswrapper[4808]: I1124 18:34:06.133361 4808 scope.go:117] "RemoveContainer" containerID="71bdc92f32291232187da14655de079d7bce31d4f354b1c4ee44415bbcba1bf2" Nov 24 18:34:06 crc kubenswrapper[4808]: E1124 18:34:06.133677 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71bdc92f32291232187da14655de079d7bce31d4f354b1c4ee44415bbcba1bf2\": container with ID starting with 71bdc92f32291232187da14655de079d7bce31d4f354b1c4ee44415bbcba1bf2 not found: ID does not exist" containerID="71bdc92f32291232187da14655de079d7bce31d4f354b1c4ee44415bbcba1bf2" Nov 24 18:34:06 crc kubenswrapper[4808]: I1124 18:34:06.133779 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71bdc92f32291232187da14655de079d7bce31d4f354b1c4ee44415bbcba1bf2"} err="failed to get container status \"71bdc92f32291232187da14655de079d7bce31d4f354b1c4ee44415bbcba1bf2\": rpc error: code = NotFound desc = could not find container \"71bdc92f32291232187da14655de079d7bce31d4f354b1c4ee44415bbcba1bf2\": container with ID starting with 71bdc92f32291232187da14655de079d7bce31d4f354b1c4ee44415bbcba1bf2 not found: ID does not exist" Nov 24 18:34:06 crc kubenswrapper[4808]: I1124 18:34:06.176449 4808 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/626e8db0-9921-43ac-a1e1-55ff96937a43-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 18:34:06 crc kubenswrapper[4808]: I1124 18:34:06.358269 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="626e8db0-9921-43ac-a1e1-55ff96937a43" path="/var/lib/kubelet/pods/626e8db0-9921-43ac-a1e1-55ff96937a43/volumes" Nov 24 18:35:55 crc kubenswrapper[4808]: I1124 18:35:55.613705 4808 scope.go:117] "RemoveContainer" containerID="39518ab101eb2177631df8a3e019cbb9530fd34f620b752bd70e5791befab7ae" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.258770 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hxddn"] Nov 24 18:36:00 crc kubenswrapper[4808]: E1124 18:36:00.259987 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="626e8db0-9921-43ac-a1e1-55ff96937a43" containerName="gather" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.260012 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="626e8db0-9921-43ac-a1e1-55ff96937a43" containerName="gather" Nov 24 18:36:00 crc kubenswrapper[4808]: E1124 18:36:00.260315 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="626e8db0-9921-43ac-a1e1-55ff96937a43" containerName="copy" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.260332 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="626e8db0-9921-43ac-a1e1-55ff96937a43" containerName="copy" Nov 24 18:36:00 crc kubenswrapper[4808]: E1124 18:36:00.260353 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4503f95-37ef-4564-9038-895faaa8d07d" containerName="extract-content" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.260366 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4503f95-37ef-4564-9038-895faaa8d07d" containerName="extract-content" Nov 24 18:36:00 crc kubenswrapper[4808]: E1124 18:36:00.260391 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4503f95-37ef-4564-9038-895faaa8d07d" containerName="registry-server" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.260403 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4503f95-37ef-4564-9038-895faaa8d07d" containerName="registry-server" Nov 24 18:36:00 crc kubenswrapper[4808]: E1124 18:36:00.260434 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4503f95-37ef-4564-9038-895faaa8d07d" containerName="extract-utilities" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.260448 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4503f95-37ef-4564-9038-895faaa8d07d" containerName="extract-utilities" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.260857 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="626e8db0-9921-43ac-a1e1-55ff96937a43" containerName="gather" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.260888 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4503f95-37ef-4564-9038-895faaa8d07d" containerName="registry-server" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.260942 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="626e8db0-9921-43ac-a1e1-55ff96937a43" containerName="copy" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.263567 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.292215 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hxddn"] Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.347069 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4829e927-eea9-4df3-8585-9f866708431f-catalog-content\") pod \"redhat-marketplace-hxddn\" (UID: \"4829e927-eea9-4df3-8585-9f866708431f\") " pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.347133 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9l56\" (UniqueName: \"kubernetes.io/projected/4829e927-eea9-4df3-8585-9f866708431f-kube-api-access-f9l56\") pod \"redhat-marketplace-hxddn\" (UID: \"4829e927-eea9-4df3-8585-9f866708431f\") " pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.347645 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4829e927-eea9-4df3-8585-9f866708431f-utilities\") pod \"redhat-marketplace-hxddn\" (UID: \"4829e927-eea9-4df3-8585-9f866708431f\") " pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.449953 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4829e927-eea9-4df3-8585-9f866708431f-catalog-content\") pod \"redhat-marketplace-hxddn\" (UID: \"4829e927-eea9-4df3-8585-9f866708431f\") " pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.450040 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9l56\" (UniqueName: \"kubernetes.io/projected/4829e927-eea9-4df3-8585-9f866708431f-kube-api-access-f9l56\") pod \"redhat-marketplace-hxddn\" (UID: \"4829e927-eea9-4df3-8585-9f866708431f\") " pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.450553 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4829e927-eea9-4df3-8585-9f866708431f-catalog-content\") pod \"redhat-marketplace-hxddn\" (UID: \"4829e927-eea9-4df3-8585-9f866708431f\") " pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.451307 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4829e927-eea9-4df3-8585-9f866708431f-utilities\") pod \"redhat-marketplace-hxddn\" (UID: \"4829e927-eea9-4df3-8585-9f866708431f\") " pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.451668 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4829e927-eea9-4df3-8585-9f866708431f-utilities\") pod \"redhat-marketplace-hxddn\" (UID: \"4829e927-eea9-4df3-8585-9f866708431f\") " pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.476550 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9l56\" (UniqueName: \"kubernetes.io/projected/4829e927-eea9-4df3-8585-9f866708431f-kube-api-access-f9l56\") pod \"redhat-marketplace-hxddn\" (UID: \"4829e927-eea9-4df3-8585-9f866708431f\") " pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:00 crc kubenswrapper[4808]: I1124 18:36:00.594526 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:01 crc kubenswrapper[4808]: I1124 18:36:01.058872 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hxddn"] Nov 24 18:36:01 crc kubenswrapper[4808]: I1124 18:36:01.556258 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxddn" event={"ID":"4829e927-eea9-4df3-8585-9f866708431f","Type":"ContainerStarted","Data":"58f46612d93b4759d8a2393f94595850013cde4316ed8843b08bf41a04270a4d"} Nov 24 18:36:02 crc kubenswrapper[4808]: I1124 18:36:02.572237 4808 generic.go:334] "Generic (PLEG): container finished" podID="4829e927-eea9-4df3-8585-9f866708431f" containerID="cd16fa5e622411c8cdbd4e715ca262d5717e3c520516b1ef59f7353608d8374c" exitCode=0 Nov 24 18:36:02 crc kubenswrapper[4808]: I1124 18:36:02.572305 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxddn" event={"ID":"4829e927-eea9-4df3-8585-9f866708431f","Type":"ContainerDied","Data":"cd16fa5e622411c8cdbd4e715ca262d5717e3c520516b1ef59f7353608d8374c"} Nov 24 18:36:04 crc kubenswrapper[4808]: I1124 18:36:04.594485 4808 generic.go:334] "Generic (PLEG): container finished" podID="4829e927-eea9-4df3-8585-9f866708431f" containerID="019963e587577d495a35a4908337d0f4baa5cad1da03c3123e83676824ab2679" exitCode=0 Nov 24 18:36:04 crc kubenswrapper[4808]: I1124 18:36:04.594600 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxddn" event={"ID":"4829e927-eea9-4df3-8585-9f866708431f","Type":"ContainerDied","Data":"019963e587577d495a35a4908337d0f4baa5cad1da03c3123e83676824ab2679"} Nov 24 18:36:05 crc kubenswrapper[4808]: I1124 18:36:05.605381 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxddn" event={"ID":"4829e927-eea9-4df3-8585-9f866708431f","Type":"ContainerStarted","Data":"7fa2426990fb6034db8b57ce92c7b348af950762969b3189ff99c040f53feb05"} Nov 24 18:36:05 crc kubenswrapper[4808]: I1124 18:36:05.626309 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hxddn" podStartSLOduration=3.02218759 podStartE2EDuration="5.626290942s" podCreationTimestamp="2025-11-24 18:36:00 +0000 UTC" firstStartedPulling="2025-11-24 18:36:02.575635167 +0000 UTC m=+4155.173302999" lastFinishedPulling="2025-11-24 18:36:05.179738549 +0000 UTC m=+4157.777406351" observedRunningTime="2025-11-24 18:36:05.621247366 +0000 UTC m=+4158.218915188" watchObservedRunningTime="2025-11-24 18:36:05.626290942 +0000 UTC m=+4158.223958744" Nov 24 18:36:06 crc kubenswrapper[4808]: I1124 18:36:06.522711 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:36:06 crc kubenswrapper[4808]: I1124 18:36:06.523044 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:36:10 crc kubenswrapper[4808]: I1124 18:36:10.594919 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:10 crc kubenswrapper[4808]: I1124 18:36:10.596836 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:10 crc kubenswrapper[4808]: I1124 18:36:10.687101 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:11 crc kubenswrapper[4808]: I1124 18:36:11.743120 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:11 crc kubenswrapper[4808]: I1124 18:36:11.811543 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hxddn"] Nov 24 18:36:13 crc kubenswrapper[4808]: I1124 18:36:13.691110 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hxddn" podUID="4829e927-eea9-4df3-8585-9f866708431f" containerName="registry-server" containerID="cri-o://7fa2426990fb6034db8b57ce92c7b348af950762969b3189ff99c040f53feb05" gracePeriod=2 Nov 24 18:36:14 crc kubenswrapper[4808]: I1124 18:36:14.705775 4808 generic.go:334] "Generic (PLEG): container finished" podID="4829e927-eea9-4df3-8585-9f866708431f" containerID="7fa2426990fb6034db8b57ce92c7b348af950762969b3189ff99c040f53feb05" exitCode=0 Nov 24 18:36:14 crc kubenswrapper[4808]: I1124 18:36:14.705875 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxddn" event={"ID":"4829e927-eea9-4df3-8585-9f866708431f","Type":"ContainerDied","Data":"7fa2426990fb6034db8b57ce92c7b348af950762969b3189ff99c040f53feb05"} Nov 24 18:36:14 crc kubenswrapper[4808]: I1124 18:36:14.863397 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:14 crc kubenswrapper[4808]: I1124 18:36:14.922097 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4829e927-eea9-4df3-8585-9f866708431f-catalog-content\") pod \"4829e927-eea9-4df3-8585-9f866708431f\" (UID: \"4829e927-eea9-4df3-8585-9f866708431f\") " Nov 24 18:36:14 crc kubenswrapper[4808]: I1124 18:36:14.922209 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9l56\" (UniqueName: \"kubernetes.io/projected/4829e927-eea9-4df3-8585-9f866708431f-kube-api-access-f9l56\") pod \"4829e927-eea9-4df3-8585-9f866708431f\" (UID: \"4829e927-eea9-4df3-8585-9f866708431f\") " Nov 24 18:36:14 crc kubenswrapper[4808]: I1124 18:36:14.922248 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4829e927-eea9-4df3-8585-9f866708431f-utilities\") pod \"4829e927-eea9-4df3-8585-9f866708431f\" (UID: \"4829e927-eea9-4df3-8585-9f866708431f\") " Nov 24 18:36:14 crc kubenswrapper[4808]: I1124 18:36:14.924135 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4829e927-eea9-4df3-8585-9f866708431f-utilities" (OuterVolumeSpecName: "utilities") pod "4829e927-eea9-4df3-8585-9f866708431f" (UID: "4829e927-eea9-4df3-8585-9f866708431f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:36:14 crc kubenswrapper[4808]: I1124 18:36:14.943770 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4829e927-eea9-4df3-8585-9f866708431f-kube-api-access-f9l56" (OuterVolumeSpecName: "kube-api-access-f9l56") pod "4829e927-eea9-4df3-8585-9f866708431f" (UID: "4829e927-eea9-4df3-8585-9f866708431f"). InnerVolumeSpecName "kube-api-access-f9l56". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:36:14 crc kubenswrapper[4808]: I1124 18:36:14.955094 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4829e927-eea9-4df3-8585-9f866708431f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4829e927-eea9-4df3-8585-9f866708431f" (UID: "4829e927-eea9-4df3-8585-9f866708431f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:36:15 crc kubenswrapper[4808]: I1124 18:36:15.023850 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9l56\" (UniqueName: \"kubernetes.io/projected/4829e927-eea9-4df3-8585-9f866708431f-kube-api-access-f9l56\") on node \"crc\" DevicePath \"\"" Nov 24 18:36:15 crc kubenswrapper[4808]: I1124 18:36:15.023880 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4829e927-eea9-4df3-8585-9f866708431f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:36:15 crc kubenswrapper[4808]: I1124 18:36:15.023892 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4829e927-eea9-4df3-8585-9f866708431f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:36:15 crc kubenswrapper[4808]: I1124 18:36:15.754561 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxddn" event={"ID":"4829e927-eea9-4df3-8585-9f866708431f","Type":"ContainerDied","Data":"58f46612d93b4759d8a2393f94595850013cde4316ed8843b08bf41a04270a4d"} Nov 24 18:36:15 crc kubenswrapper[4808]: I1124 18:36:15.754659 4808 scope.go:117] "RemoveContainer" containerID="7fa2426990fb6034db8b57ce92c7b348af950762969b3189ff99c040f53feb05" Nov 24 18:36:15 crc kubenswrapper[4808]: I1124 18:36:15.754679 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hxddn" Nov 24 18:36:15 crc kubenswrapper[4808]: I1124 18:36:15.792055 4808 scope.go:117] "RemoveContainer" containerID="019963e587577d495a35a4908337d0f4baa5cad1da03c3123e83676824ab2679" Nov 24 18:36:15 crc kubenswrapper[4808]: I1124 18:36:15.809481 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hxddn"] Nov 24 18:36:15 crc kubenswrapper[4808]: I1124 18:36:15.821803 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hxddn"] Nov 24 18:36:15 crc kubenswrapper[4808]: I1124 18:36:15.828509 4808 scope.go:117] "RemoveContainer" containerID="cd16fa5e622411c8cdbd4e715ca262d5717e3c520516b1ef59f7353608d8374c" Nov 24 18:36:16 crc kubenswrapper[4808]: I1124 18:36:16.364281 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4829e927-eea9-4df3-8585-9f866708431f" path="/var/lib/kubelet/pods/4829e927-eea9-4df3-8585-9f866708431f/volumes" Nov 24 18:36:36 crc kubenswrapper[4808]: I1124 18:36:36.525407 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:36:36 crc kubenswrapper[4808]: I1124 18:36:36.526357 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:37:06 crc kubenswrapper[4808]: I1124 18:37:06.523289 4808 patch_prober.go:28] interesting pod/machine-config-daemon-ht87w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:37:06 crc kubenswrapper[4808]: I1124 18:37:06.524208 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:37:06 crc kubenswrapper[4808]: I1124 18:37:06.524255 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" Nov 24 18:37:06 crc kubenswrapper[4808]: I1124 18:37:06.524949 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"edaf6fa37eb97fe11a3c54d186e8fc013d121dd3e78f8413570dc6ca074c8458"} pod="openshift-machine-config-operator/machine-config-daemon-ht87w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 18:37:06 crc kubenswrapper[4808]: I1124 18:37:06.524991 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" podUID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerName="machine-config-daemon" containerID="cri-o://edaf6fa37eb97fe11a3c54d186e8fc013d121dd3e78f8413570dc6ca074c8458" gracePeriod=600 Nov 24 18:37:07 crc kubenswrapper[4808]: I1124 18:37:07.423196 4808 generic.go:334] "Generic (PLEG): container finished" podID="5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0" containerID="edaf6fa37eb97fe11a3c54d186e8fc013d121dd3e78f8413570dc6ca074c8458" exitCode=0 Nov 24 18:37:07 crc kubenswrapper[4808]: I1124 18:37:07.423280 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerDied","Data":"edaf6fa37eb97fe11a3c54d186e8fc013d121dd3e78f8413570dc6ca074c8458"} Nov 24 18:37:07 crc kubenswrapper[4808]: I1124 18:37:07.423641 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ht87w" event={"ID":"5b86a0e3-4cb3-4ca4-a9ac-c66e137415d0","Type":"ContainerStarted","Data":"6737566a20488a0d1e40b9d14dc811f3043bfa013f6af379302a5f2454678e2b"} Nov 24 18:37:07 crc kubenswrapper[4808]: I1124 18:37:07.423675 4808 scope.go:117] "RemoveContainer" containerID="65f6c57491316638eaf1f6bc00b242f5650566925f9664e76af3186453cce395" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.492875 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tlmwr"] Nov 24 18:38:09 crc kubenswrapper[4808]: E1124 18:38:09.494490 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4829e927-eea9-4df3-8585-9f866708431f" containerName="extract-content" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.494524 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4829e927-eea9-4df3-8585-9f866708431f" containerName="extract-content" Nov 24 18:38:09 crc kubenswrapper[4808]: E1124 18:38:09.494580 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4829e927-eea9-4df3-8585-9f866708431f" containerName="registry-server" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.494599 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4829e927-eea9-4df3-8585-9f866708431f" containerName="registry-server" Nov 24 18:38:09 crc kubenswrapper[4808]: E1124 18:38:09.494639 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4829e927-eea9-4df3-8585-9f866708431f" containerName="extract-utilities" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.494654 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="4829e927-eea9-4df3-8585-9f866708431f" containerName="extract-utilities" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.495168 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="4829e927-eea9-4df3-8585-9f866708431f" containerName="registry-server" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.498494 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tlmwr" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.524634 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tlmwr"] Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.695591 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb-catalog-content\") pod \"certified-operators-tlmwr\" (UID: \"f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb\") " pod="openshift-marketplace/certified-operators-tlmwr" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.695705 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wstm2\" (UniqueName: \"kubernetes.io/projected/f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb-kube-api-access-wstm2\") pod \"certified-operators-tlmwr\" (UID: \"f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb\") " pod="openshift-marketplace/certified-operators-tlmwr" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.695745 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb-utilities\") pod \"certified-operators-tlmwr\" (UID: \"f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb\") " pod="openshift-marketplace/certified-operators-tlmwr" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.797600 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wstm2\" (UniqueName: \"kubernetes.io/projected/f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb-kube-api-access-wstm2\") pod \"certified-operators-tlmwr\" (UID: \"f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb\") " pod="openshift-marketplace/certified-operators-tlmwr" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.797852 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb-utilities\") pod \"certified-operators-tlmwr\" (UID: \"f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb\") " pod="openshift-marketplace/certified-operators-tlmwr" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.798036 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb-catalog-content\") pod \"certified-operators-tlmwr\" (UID: \"f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb\") " pod="openshift-marketplace/certified-operators-tlmwr" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.798379 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb-utilities\") pod \"certified-operators-tlmwr\" (UID: \"f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb\") " pod="openshift-marketplace/certified-operators-tlmwr" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.798486 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb-catalog-content\") pod \"certified-operators-tlmwr\" (UID: \"f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb\") " pod="openshift-marketplace/certified-operators-tlmwr" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.825757 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wstm2\" (UniqueName: \"kubernetes.io/projected/f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb-kube-api-access-wstm2\") pod \"certified-operators-tlmwr\" (UID: \"f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb\") " pod="openshift-marketplace/certified-operators-tlmwr" Nov 24 18:38:09 crc kubenswrapper[4808]: I1124 18:38:09.839199 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tlmwr" Nov 24 18:38:10 crc kubenswrapper[4808]: I1124 18:38:10.326654 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tlmwr"] Nov 24 18:38:11 crc kubenswrapper[4808]: I1124 18:38:11.211886 4808 generic.go:334] "Generic (PLEG): container finished" podID="f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb" containerID="3fd8418f415e0a6a62e00bd436ea6dee9db2c81e5b0e52edf9bb395a76bb983a" exitCode=0 Nov 24 18:38:11 crc kubenswrapper[4808]: I1124 18:38:11.212394 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tlmwr" event={"ID":"f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb","Type":"ContainerDied","Data":"3fd8418f415e0a6a62e00bd436ea6dee9db2c81e5b0e52edf9bb395a76bb983a"} Nov 24 18:38:11 crc kubenswrapper[4808]: I1124 18:38:11.212428 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tlmwr" event={"ID":"f36cc7dc-1dde-43d6-8ef4-f51d3a2103eb","Type":"ContainerStarted","Data":"db8dda8819cf0cf38fcfd6750e4ee67afd7da5afcd87810dc739a21e8969aef8"} Nov 24 18:38:11 crc kubenswrapper[4808]: I1124 18:38:11.215319 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111123031024431 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111123032017347 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111112211016467 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111112212015440 5ustar corecore